Zipf1 All Model Test Data Result Combined

Result

Model Summaries

Model Better than base % of the times
LR_10[cache_size=0.001,treshold=0.3] 0
LR_10[cache_size=0.001,treshold=0.5] 0
LR_10[cache_size=0.001,treshold=0.6] 0
LR_10[cache_size=0.001,treshold=0.7] 0
LR_10[cache_size=0.001,treshold=0.8] 0
LR_10[cache_size=0.001,treshold=0.9] 100
LR_10[cache_size=All,treshold=0.3] 0
LR_10[cache_size=All,treshold=0.5] 0
LR_10[cache_size=All,treshold=0.6] 0
LR_10[cache_size=All,treshold=0.7] 0
LR_10[cache_size=All,treshold=0.8] 0
LR_10[cache_size=All,treshold=0.9] 0
LR_11[cache_size=0.001,treshold=0.3] 0
LR_11[cache_size=0.001,treshold=0.5] 0
LR_11[cache_size=0.001,treshold=0.6] 0
LR_11[cache_size=0.001,treshold=0.7] 0
LR_11[cache_size=0.001,treshold=0.8] 0
LR_11[cache_size=0.001,treshold=0.9] 100
LR_11[cache_size=All,treshold=0.3] 0
LR_11[cache_size=All,treshold=0.5] 0
LR_11[cache_size=All,treshold=0.6] 0
LR_11[cache_size=All,treshold=0.7] 0
LR_11[cache_size=All,treshold=0.8] 0
LR_11[cache_size=All,treshold=0.9] 0
LR_12[cache_size=0.001,treshold=0.3] 0
LR_12[cache_size=0.001,treshold=0.5] 0
LR_12[cache_size=0.001,treshold=0.6] 0
LR_12[cache_size=0.001,treshold=0.7] 0
LR_12[cache_size=0.001,treshold=0.8] 0
LR_12[cache_size=0.001,treshold=0.9] 100
LR_12[cache_size=All,treshold=0.3] 0
LR_12[cache_size=All,treshold=0.5] 0
LR_12[cache_size=All,treshold=0.6] 0
LR_12[cache_size=All,treshold=0.7] 0
LR_12[cache_size=All,treshold=0.8] 0
LR_12[cache_size=All,treshold=0.9] 0
LR_13[cache_size=0.001,treshold=0.3] 0
LR_13[cache_size=0.001,treshold=0.5] 0
LR_13[cache_size=0.001,treshold=0.6] 0
LR_13[cache_size=0.001,treshold=0.7] 0
LR_13[cache_size=0.001,treshold=0.8] 0
LR_13[cache_size=0.001,treshold=0.9] 0
LR_13[cache_size=All,treshold=0.3] 0
LR_13[cache_size=All,treshold=0.5] 20
LR_13[cache_size=All,treshold=0.6] 0
LR_13[cache_size=All,treshold=0.7] 0
LR_13[cache_size=All,treshold=0.8] 0
LR_13[cache_size=All,treshold=0.9] 0
LR_14[cache_size=0.001,treshold=0.3] 0
LR_14[cache_size=0.001,treshold=0.5] 0
LR_14[cache_size=0.001,treshold=0.6] 0
LR_14[cache_size=0.001,treshold=0.7] 0
LR_14[cache_size=0.001,treshold=0.8] 0
LR_14[cache_size=0.001,treshold=0.9] 0
LR_14[cache_size=All,treshold=0.3] 0
LR_14[cache_size=All,treshold=0.5] 20
LR_14[cache_size=All,treshold=0.6] 0
LR_14[cache_size=All,treshold=0.7] 0
LR_14[cache_size=All,treshold=0.8] 0
LR_14[cache_size=All,treshold=0.9] 0
LR_15[cache_size=0.001,treshold=0.3] 0
LR_15[cache_size=0.001,treshold=0.5] 0
LR_15[cache_size=0.001,treshold=0.6] 0
LR_15[cache_size=0.001,treshold=0.7] 0
LR_15[cache_size=0.001,treshold=0.8] 0
LR_15[cache_size=0.001,treshold=0.9] 0
LR_15[cache_size=All,treshold=0.3] 0
LR_15[cache_size=All,treshold=0.5] 20
LR_15[cache_size=All,treshold=0.6] 0
LR_15[cache_size=All,treshold=0.7] 0
LR_15[cache_size=All,treshold=0.8] 0
LR_15[cache_size=All,treshold=0.9] 0
LR_1[cache_size=0.001,treshold=0.5] 0
LR_1[cache_size=All,treshold=0.5] 0
LR_1_log[cache_size=0.001,treshold=0.5] 0
LR_1_log[cache_size=All,treshold=0.5] 0
LR_1_mean[cache_size=0.001,treshold=0.5] 0
LR_1_mean[cache_size=All,treshold=0.5] 20
LR_1_robust_scaler[cache_size=0.001,treshold=0.5] 0
LR_1_robust_scaler[cache_size=All,treshold=0.5] 0
LR_1_std_scaler[cache_size=0.001,treshold=0.5] 0
LR_1_std_scaler[cache_size=All,treshold=0.5] 0
LR_2[cache_size=0.001,treshold=0.5] 0
LR_2[cache_size=All,treshold=0.5] 0
LR_2_log[cache_size=0.001,treshold=0.5] 0
LR_2_log[cache_size=All,treshold=0.5] 0
LR_2_mean[cache_size=0.001,treshold=0.5] 0
LR_2_mean[cache_size=All,treshold=0.5] 20
LR_3[cache_size=0.001,treshold=0.5] 0
LR_3[cache_size=All,treshold=0.5] 0
LR_3_log[cache_size=0.001,treshold=0.5] 0
LR_3_log[cache_size=All,treshold=0.5] 0
LR_3_mean[cache_size=0.001,treshold=0.5] 0
LR_3_mean[cache_size=All,treshold=0.5] 0
LR_4[cache_size=0.001,treshold=0.5] 0
LR_4[cache_size=All,treshold=0.5] 0
LR_4_log[cache_size=0.001,treshold=0.5] 0
LR_4_log[cache_size=All,treshold=0.5] 0
LR_4_mean[cache_size=0.001,treshold=0.5] 0
LR_4_mean[cache_size=All,treshold=0.5] 0
LR_4_robust_scaler[cache_size=0.001,treshold=0.5] 0
LR_4_robust_scaler[cache_size=All,treshold=0.5] 0
LR_4_std_scaler[cache_size=0.001,treshold=0.5] 0
LR_4_std_scaler[cache_size=All,treshold=0.5] 0
LR_5[cache_size=0.001,treshold=0.5] 0
LR_5[cache_size=All,treshold=0.5] 0
LR_5_imba[cache_size=0.001,treshold=0.5] 0
LR_5_imba[cache_size=All,treshold=0.5] 0
LR_6[cache_size=0.001,treshold=0.5] 0
LR_6[cache_size=All,treshold=0.5] 0
LR_6_imba[cache_size=0.001,treshold=0.5] 0
LR_6_imba[cache_size=All,treshold=0.5] 0
LR_7[cache_size=0.001,treshold=0.3] 0
LR_7[cache_size=0.001,treshold=0.5] 0
LR_7[cache_size=0.001,treshold=0.6] 0
LR_7[cache_size=0.001,treshold=0.7] 0
LR_7[cache_size=0.001,treshold=0.8] 0
LR_7[cache_size=0.001,treshold=0.9] 100
LR_7[cache_size=All,treshold=0.3] 0
LR_7[cache_size=All,treshold=0.5] 0
LR_7[cache_size=All,treshold=0.6] 0
LR_7[cache_size=All,treshold=0.7] 0
LR_7[cache_size=All,treshold=0.8] 0
LR_7[cache_size=All,treshold=0.9] 0
LR_8[cache_size=0.001,treshold=0.3] 0
LR_8[cache_size=0.001,treshold=0.5] 0
LR_8[cache_size=0.001,treshold=0.6] 0
LR_8[cache_size=0.001,treshold=0.7] 0
LR_8[cache_size=0.001,treshold=0.8] 0
LR_8[cache_size=0.001,treshold=0.9] 100
LR_8[cache_size=All,treshold=0.3] 0
LR_8[cache_size=All,treshold=0.5] 0
LR_8[cache_size=All,treshold=0.6] 0
LR_8[cache_size=All,treshold=0.7] 0
LR_8[cache_size=All,treshold=0.8] 0
LR_8[cache_size=All,treshold=0.9] 0
LR_9[cache_size=0.001,treshold=0.3] 0
LR_9[cache_size=0.001,treshold=0.5] 0
LR_9[cache_size=0.001,treshold=0.6] 0
LR_9[cache_size=0.001,treshold=0.7] 0
LR_9[cache_size=0.001,treshold=0.8] 0
LR_9[cache_size=0.001,treshold=0.9] 100
LR_9[cache_size=All,treshold=0.3] 0
LR_9[cache_size=All,treshold=0.5] 0
LR_9[cache_size=All,treshold=0.6] 0
LR_9[cache_size=All,treshold=0.7] 0
LR_9[cache_size=All,treshold=0.8] 0
LR_9[cache_size=All,treshold=0.9] 0
LR_10[cache_size=0.01,treshold=0.3] 0
LR_10[cache_size=0.01,treshold=0.5] 0
LR_10[cache_size=0.01,treshold=0.6] 0
LR_10[cache_size=0.01,treshold=0.7] 0
LR_10[cache_size=0.01,treshold=0.8] 0
LR_10[cache_size=0.01,treshold=0.9] 0
LR_11[cache_size=0.01,treshold=0.3] 0
LR_11[cache_size=0.01,treshold=0.5] 0
LR_11[cache_size=0.01,treshold=0.6] 0
LR_11[cache_size=0.01,treshold=0.7] 0
LR_11[cache_size=0.01,treshold=0.8] 0
LR_11[cache_size=0.01,treshold=0.9] 0
LR_12[cache_size=0.01,treshold=0.3] 0
LR_12[cache_size=0.01,treshold=0.5] 0
LR_12[cache_size=0.01,treshold=0.6] 0
LR_12[cache_size=0.01,treshold=0.7] 0
LR_12[cache_size=0.01,treshold=0.8] 0
LR_12[cache_size=0.01,treshold=0.9] 0
LR_13[cache_size=0.01,treshold=0.3] 0
LR_13[cache_size=0.01,treshold=0.5] 0
LR_13[cache_size=0.01,treshold=0.6] 0
LR_13[cache_size=0.01,treshold=0.7] 0
LR_13[cache_size=0.01,treshold=0.8] 0
LR_13[cache_size=0.01,treshold=0.9] 0
LR_14[cache_size=0.01,treshold=0.3] 0
LR_14[cache_size=0.01,treshold=0.5] 0
LR_14[cache_size=0.01,treshold=0.6] 0
LR_14[cache_size=0.01,treshold=0.7] 0
LR_14[cache_size=0.01,treshold=0.8] 0
LR_14[cache_size=0.01,treshold=0.9] 0
LR_15[cache_size=0.01,treshold=0.3] 0
LR_15[cache_size=0.01,treshold=0.5] 0
LR_15[cache_size=0.01,treshold=0.6] 0
LR_15[cache_size=0.01,treshold=0.7] 0
LR_15[cache_size=0.01,treshold=0.8] 0
LR_15[cache_size=0.01,treshold=0.9] 0
LR_1[cache_size=0.01,treshold=0.5] 0
LR_1_log[cache_size=0.01,treshold=0.5] 0
LR_1_mean[cache_size=0.01,treshold=0.5] 0
LR_1_robust_scaler[cache_size=0.01,treshold=0.5] 0
LR_1_std_scaler[cache_size=0.01,treshold=0.5] 0
LR_2[cache_size=0.01,treshold=0.5] 0
LR_2_log[cache_size=0.01,treshold=0.5] 0
LR_2_mean[cache_size=0.01,treshold=0.5] 0
LR_3[cache_size=0.01,treshold=0.5] 0
LR_3_log[cache_size=0.01,treshold=0.5] 0
LR_3_mean[cache_size=0.01,treshold=0.5] 0
LR_4[cache_size=0.01,treshold=0.5] 0
LR_4_log[cache_size=0.01,treshold=0.5] 0
LR_4_mean[cache_size=0.01,treshold=0.5] 0
LR_4_robust_scaler[cache_size=0.01,treshold=0.5] 0
LR_4_std_scaler[cache_size=0.01,treshold=0.5] 0
LR_5[cache_size=0.01,treshold=0.5] 0
LR_5_imba[cache_size=0.01,treshold=0.5] 0
LR_6[cache_size=0.01,treshold=0.5] 0
LR_6_imba[cache_size=0.01,treshold=0.5] 0
LR_7[cache_size=0.01,treshold=0.3] 0
LR_7[cache_size=0.01,treshold=0.5] 0
LR_7[cache_size=0.01,treshold=0.6] 0
LR_7[cache_size=0.01,treshold=0.7] 0
LR_7[cache_size=0.01,treshold=0.8] 0
LR_7[cache_size=0.01,treshold=0.9] 100
LR_8[cache_size=0.01,treshold=0.3] 0
LR_8[cache_size=0.01,treshold=0.5] 0
LR_8[cache_size=0.01,treshold=0.6] 0
LR_8[cache_size=0.01,treshold=0.7] 0
LR_8[cache_size=0.01,treshold=0.8] 0
LR_8[cache_size=0.01,treshold=0.9] 100
LR_9[cache_size=0.01,treshold=0.3] 0
LR_9[cache_size=0.01,treshold=0.5] 0
LR_9[cache_size=0.01,treshold=0.6] 0
LR_9[cache_size=0.01,treshold=0.7] 0
LR_9[cache_size=0.01,treshold=0.8] 0
LR_9[cache_size=0.01,treshold=0.9] 100
LR_10[cache_size=0.1,treshold=0.3] 0
LR_10[cache_size=0.1,treshold=0.5] 0
LR_10[cache_size=0.1,treshold=0.6] 0
LR_10[cache_size=0.1,treshold=0.7] 0
LR_10[cache_size=0.1,treshold=0.8] 0
LR_10[cache_size=0.1,treshold=0.9] 0
LR_11[cache_size=0.1,treshold=0.3] 0
LR_11[cache_size=0.1,treshold=0.5] 0
LR_11[cache_size=0.1,treshold=0.6] 0
LR_11[cache_size=0.1,treshold=0.7] 0
LR_11[cache_size=0.1,treshold=0.8] 0
LR_11[cache_size=0.1,treshold=0.9] 0
LR_12[cache_size=0.1,treshold=0.3] 0
LR_12[cache_size=0.1,treshold=0.5] 0
LR_12[cache_size=0.1,treshold=0.6] 0
LR_12[cache_size=0.1,treshold=0.7] 0
LR_12[cache_size=0.1,treshold=0.8] 0
LR_12[cache_size=0.1,treshold=0.9] 0
LR_13[cache_size=0.1,treshold=0.3] 0
LR_13[cache_size=0.1,treshold=0.5] 0
LR_13[cache_size=0.1,treshold=0.6] 0
LR_13[cache_size=0.1,treshold=0.7] 0
LR_13[cache_size=0.1,treshold=0.8] 0
LR_13[cache_size=0.1,treshold=0.9] 0
LR_14[cache_size=0.1,treshold=0.3] 0
LR_14[cache_size=0.1,treshold=0.5] 0
LR_14[cache_size=0.1,treshold=0.6] 0
LR_14[cache_size=0.1,treshold=0.7] 0
LR_14[cache_size=0.1,treshold=0.8] 0
LR_14[cache_size=0.1,treshold=0.9] 0
LR_15[cache_size=0.1,treshold=0.3] 0
LR_15[cache_size=0.1,treshold=0.5] 0
LR_15[cache_size=0.1,treshold=0.6] 0
LR_15[cache_size=0.1,treshold=0.7] 0
LR_15[cache_size=0.1,treshold=0.8] 0
LR_15[cache_size=0.1,treshold=0.9] 0
LR_1[cache_size=0.1,treshold=0.5] 0
LR_1_log[cache_size=0.1,treshold=0.5] 0
LR_1_mean[cache_size=0.1,treshold=0.5] 0
LR_1_robust_scaler[cache_size=0.1,treshold=0.5] 0
LR_1_std_scaler[cache_size=0.1,treshold=0.5] 0
LR_2[cache_size=0.1,treshold=0.5] 0
LR_2_log[cache_size=0.1,treshold=0.5] 0
LR_2_mean[cache_size=0.1,treshold=0.5] 0
LR_3[cache_size=0.1,treshold=0.5] 0
LR_3_log[cache_size=0.1,treshold=0.5] 0
LR_3_mean[cache_size=0.1,treshold=0.5] 0
LR_4[cache_size=0.1,treshold=0.5] 0
LR_4_log[cache_size=0.1,treshold=0.5] 0
LR_4_mean[cache_size=0.1,treshold=0.5] 0
LR_4_robust_scaler[cache_size=0.1,treshold=0.5] 0
LR_4_std_scaler[cache_size=0.1,treshold=0.5] 0
LR_5[cache_size=0.1,treshold=0.5] 0
LR_5_imba[cache_size=0.1,treshold=0.5] 0
LR_6[cache_size=0.1,treshold=0.5] 0
LR_6_imba[cache_size=0.1,treshold=0.5] 0
LR_7[cache_size=0.1,treshold=0.3] 0
LR_7[cache_size=0.1,treshold=0.5] 0
LR_7[cache_size=0.1,treshold=0.6] 0
LR_7[cache_size=0.1,treshold=0.7] 0
LR_7[cache_size=0.1,treshold=0.8] 0
LR_7[cache_size=0.1,treshold=0.9] 0
LR_8[cache_size=0.1,treshold=0.3] 0
LR_8[cache_size=0.1,treshold=0.5] 0
LR_8[cache_size=0.1,treshold=0.6] 0
LR_8[cache_size=0.1,treshold=0.7] 0
LR_8[cache_size=0.1,treshold=0.8] 0
LR_8[cache_size=0.1,treshold=0.9] 0
LR_9[cache_size=0.1,treshold=0.3] 0
LR_9[cache_size=0.1,treshold=0.5] 0
LR_9[cache_size=0.1,treshold=0.6] 0
LR_9[cache_size=0.1,treshold=0.7] 0
LR_9[cache_size=0.1,treshold=0.8] 0
LR_9[cache_size=0.1,treshold=0.9] 0
LR_10[cache_size=0.2,treshold=0.3] 0
LR_10[cache_size=0.2,treshold=0.5] 0
LR_10[cache_size=0.2,treshold=0.6] 0
LR_10[cache_size=0.2,treshold=0.7] 0
LR_10[cache_size=0.2,treshold=0.8] 0
LR_10[cache_size=0.2,treshold=0.9] 0
LR_11[cache_size=0.2,treshold=0.3] 0
LR_11[cache_size=0.2,treshold=0.5] 0
LR_11[cache_size=0.2,treshold=0.6] 0
LR_11[cache_size=0.2,treshold=0.7] 0
LR_11[cache_size=0.2,treshold=0.8] 0
LR_11[cache_size=0.2,treshold=0.9] 0
LR_12[cache_size=0.2,treshold=0.3] 0
LR_12[cache_size=0.2,treshold=0.5] 0
LR_12[cache_size=0.2,treshold=0.6] 0
LR_12[cache_size=0.2,treshold=0.7] 0
LR_12[cache_size=0.2,treshold=0.8] 0
LR_12[cache_size=0.2,treshold=0.9] 0
LR_13[cache_size=0.2,treshold=0.3] 0
LR_13[cache_size=0.2,treshold=0.5] 0
LR_13[cache_size=0.2,treshold=0.6] 0
LR_13[cache_size=0.2,treshold=0.7] 0
LR_13[cache_size=0.2,treshold=0.8] 0
LR_13[cache_size=0.2,treshold=0.9] 0
LR_14[cache_size=0.2,treshold=0.3] 0
LR_14[cache_size=0.2,treshold=0.5] 0
LR_14[cache_size=0.2,treshold=0.6] 0
LR_14[cache_size=0.2,treshold=0.7] 0
LR_14[cache_size=0.2,treshold=0.8] 0
LR_14[cache_size=0.2,treshold=0.9] 0
LR_15[cache_size=0.2,treshold=0.3] 0
LR_15[cache_size=0.2,treshold=0.5] 0
LR_15[cache_size=0.2,treshold=0.6] 0
LR_15[cache_size=0.2,treshold=0.7] 0
LR_15[cache_size=0.2,treshold=0.8] 0
LR_15[cache_size=0.2,treshold=0.9] 0
LR_1[cache_size=0.2,treshold=0.5] 0
LR_1_log[cache_size=0.2,treshold=0.5] 0
LR_1_mean[cache_size=0.2,treshold=0.5] 0
LR_1_robust_scaler[cache_size=0.2,treshold=0.5] 0
LR_1_std_scaler[cache_size=0.2,treshold=0.5] 0
LR_2[cache_size=0.2,treshold=0.5] 0
LR_2_log[cache_size=0.2,treshold=0.5] 0
LR_2_mean[cache_size=0.2,treshold=0.5] 0
LR_3[cache_size=0.2,treshold=0.5] 0
LR_3_log[cache_size=0.2,treshold=0.5] 0
LR_3_mean[cache_size=0.2,treshold=0.5] 0
LR_4[cache_size=0.2,treshold=0.5] 0
LR_4_log[cache_size=0.2,treshold=0.5] 0
LR_4_mean[cache_size=0.2,treshold=0.5] 0
LR_4_robust_scaler[cache_size=0.2,treshold=0.5] 0
LR_4_std_scaler[cache_size=0.2,treshold=0.5] 0
LR_5[cache_size=0.2,treshold=0.5] 0
LR_5_imba[cache_size=0.2,treshold=0.5] 0
LR_6[cache_size=0.2,treshold=0.5] 0
LR_6_imba[cache_size=0.2,treshold=0.5] 0
LR_7[cache_size=0.2,treshold=0.3] 0
LR_7[cache_size=0.2,treshold=0.5] 0
LR_7[cache_size=0.2,treshold=0.6] 0
LR_7[cache_size=0.2,treshold=0.7] 0
LR_7[cache_size=0.2,treshold=0.8] 0
LR_7[cache_size=0.2,treshold=0.9] 0
LR_8[cache_size=0.2,treshold=0.3] 0
LR_8[cache_size=0.2,treshold=0.5] 0
LR_8[cache_size=0.2,treshold=0.6] 0
LR_8[cache_size=0.2,treshold=0.7] 0
LR_8[cache_size=0.2,treshold=0.8] 0
LR_8[cache_size=0.2,treshold=0.9] 0
LR_9[cache_size=0.2,treshold=0.3] 0
LR_9[cache_size=0.2,treshold=0.5] 0
LR_9[cache_size=0.2,treshold=0.6] 0
LR_9[cache_size=0.2,treshold=0.7] 0
LR_9[cache_size=0.2,treshold=0.8] 0
LR_9[cache_size=0.2,treshold=0.9] 0
LR_10[cache_size=0.4,treshold=0.3] 0
LR_10[cache_size=0.4,treshold=0.5] 0
LR_10[cache_size=0.4,treshold=0.6] 0
LR_10[cache_size=0.4,treshold=0.7] 0
LR_10[cache_size=0.4,treshold=0.8] 0
LR_10[cache_size=0.4,treshold=0.9] 0
LR_11[cache_size=0.4,treshold=0.3] 0
LR_11[cache_size=0.4,treshold=0.5] 0
LR_11[cache_size=0.4,treshold=0.6] 0
LR_11[cache_size=0.4,treshold=0.7] 0
LR_11[cache_size=0.4,treshold=0.8] 0
LR_11[cache_size=0.4,treshold=0.9] 0
LR_12[cache_size=0.4,treshold=0.3] 0
LR_12[cache_size=0.4,treshold=0.5] 0
LR_12[cache_size=0.4,treshold=0.6] 0
LR_12[cache_size=0.4,treshold=0.7] 0
LR_12[cache_size=0.4,treshold=0.8] 0
LR_12[cache_size=0.4,treshold=0.9] 0
LR_13[cache_size=0.4,treshold=0.3] 0
LR_13[cache_size=0.4,treshold=0.5] 0
LR_13[cache_size=0.4,treshold=0.6] 0
LR_13[cache_size=0.4,treshold=0.7] 0
LR_13[cache_size=0.4,treshold=0.8] 0
LR_13[cache_size=0.4,treshold=0.9] 0
LR_14[cache_size=0.4,treshold=0.3] 0
LR_14[cache_size=0.4,treshold=0.5] 0
LR_14[cache_size=0.4,treshold=0.6] 0
LR_14[cache_size=0.4,treshold=0.7] 0
LR_14[cache_size=0.4,treshold=0.8] 0
LR_14[cache_size=0.4,treshold=0.9] 0
LR_15[cache_size=0.4,treshold=0.3] 0
LR_15[cache_size=0.4,treshold=0.5] 0
LR_15[cache_size=0.4,treshold=0.6] 0
LR_15[cache_size=0.4,treshold=0.7] 0
LR_15[cache_size=0.4,treshold=0.8] 0
LR_15[cache_size=0.4,treshold=0.9] 0
LR_1[cache_size=0.4,treshold=0.5] 0
LR_1_log[cache_size=0.4,treshold=0.5] 0
LR_1_mean[cache_size=0.4,treshold=0.5] 0
LR_1_robust_scaler[cache_size=0.4,treshold=0.5] 0
LR_1_std_scaler[cache_size=0.4,treshold=0.5] 0
LR_2[cache_size=0.4,treshold=0.5] 0
LR_2_log[cache_size=0.4,treshold=0.5] 0
LR_2_mean[cache_size=0.4,treshold=0.5] 0
LR_3[cache_size=0.4,treshold=0.5] 0
LR_3_log[cache_size=0.4,treshold=0.5] 0
LR_3_mean[cache_size=0.4,treshold=0.5] 0
LR_4[cache_size=0.4,treshold=0.5] 0
LR_4_log[cache_size=0.4,treshold=0.5] 0
LR_4_mean[cache_size=0.4,treshold=0.5] 0
LR_4_robust_scaler[cache_size=0.4,treshold=0.5] 0
LR_4_std_scaler[cache_size=0.4,treshold=0.5] 0
LR_5[cache_size=0.4,treshold=0.5] 0
LR_5_imba[cache_size=0.4,treshold=0.5] 0
LR_6[cache_size=0.4,treshold=0.5] 0
LR_6_imba[cache_size=0.4,treshold=0.5] 0
LR_7[cache_size=0.4,treshold=0.3] 0
LR_7[cache_size=0.4,treshold=0.5] 0
LR_7[cache_size=0.4,treshold=0.6] 0
LR_7[cache_size=0.4,treshold=0.7] 0
LR_7[cache_size=0.4,treshold=0.8] 0
LR_7[cache_size=0.4,treshold=0.9] 0
LR_8[cache_size=0.4,treshold=0.3] 0
LR_8[cache_size=0.4,treshold=0.5] 0
LR_8[cache_size=0.4,treshold=0.6] 0
LR_8[cache_size=0.4,treshold=0.7] 0
LR_8[cache_size=0.4,treshold=0.8] 0
LR_8[cache_size=0.4,treshold=0.9] 0
LR_9[cache_size=0.4,treshold=0.3] 0
LR_9[cache_size=0.4,treshold=0.5] 0
LR_9[cache_size=0.4,treshold=0.6] 0
LR_9[cache_size=0.4,treshold=0.7] 0
LR_9[cache_size=0.4,treshold=0.8] 0
LR_9[cache_size=0.4,treshold=0.9] 0
Offline Clock 1st iteration 0
Offline Clock 2nd iteration 100
Zipf Optimal Distribution 100

Promotion Reduced (%)

Model Max Min Avg Mdn
LR_10[cache_size=0.001,treshold=0.3] 64.487 64.4568 64.4714 64.4716
LR_10[cache_size=0.001,treshold=0.5] 53.0073 52.9553 52.9846 52.9966
LR_10[cache_size=0.001,treshold=0.6] 45.4336 45.3647 45.3964 45.3967
LR_10[cache_size=0.001,treshold=0.7] 35.2239 35.1603 35.1838 35.1825
LR_10[cache_size=0.001,treshold=0.8] 20.19 20.1546 20.1658 20.1625
LR_10[cache_size=0.001,treshold=0.9] 0.0489568 0.0458181 0.0475212 0.0474412
LR_10[cache_size=All,treshold=0.3] 99.9133 65.2775 91.592 99.2999
LR_10[cache_size=All,treshold=0.5] 99.6467 8.24613 75.4498 97.1811
LR_10[cache_size=All,treshold=0.6] 0 0 0 0
LR_10[cache_size=All,treshold=0.7] 0 0 0 0
LR_10[cache_size=All,treshold=0.8] 0 0 0 0
LR_10[cache_size=All,treshold=0.9] 0 0 0 0
LR_11[cache_size=0.001,treshold=0.3] 64.9619 64.9316 64.9501 64.9501
LR_11[cache_size=0.001,treshold=0.5] 53.307 53.2581 53.2857 53.3009
LR_11[cache_size=0.001,treshold=0.6] 45.4136 45.3526 45.3804 45.3735
LR_11[cache_size=0.001,treshold=0.7] 34.6595 34.6152 34.6336 34.6296
LR_11[cache_size=0.001,treshold=0.8] 19.2793 19.241 19.2604 19.2604
LR_11[cache_size=0.001,treshold=0.9] 0.811724 0.806042 0.809589 0.809746
LR_11[cache_size=All,treshold=0.3] 99.9133 65.2759 91.5911 99.2995
LR_11[cache_size=All,treshold=0.5] 99.6478 8.28082 75.4773 97.1887
LR_11[cache_size=All,treshold=0.6] 0 0 0 0
LR_11[cache_size=All,treshold=0.7] 0 0 0 0
LR_11[cache_size=All,treshold=0.8] 0 0 0 0
LR_11[cache_size=All,treshold=0.9] 0 0 0 0
LR_12[cache_size=0.001,treshold=0.3] 64.6 64.5723 64.586 64.5864
LR_12[cache_size=0.001,treshold=0.5] 52.9772 52.9285 52.9559 52.9715
LR_12[cache_size=0.001,treshold=0.6] 45.3007 45.2448 45.2688 45.2616
LR_12[cache_size=0.001,treshold=0.7] 35.0088 34.9565 34.9749 34.9686
LR_12[cache_size=0.001,treshold=0.8] 20.1304 20.0888 20.1081 20.1093
LR_12[cache_size=0.001,treshold=0.9] 0.614198 0.608547 0.611838 0.611802
LR_12[cache_size=All,treshold=0.3] 99.9136 65.269 91.5943 99.3023
LR_12[cache_size=All,treshold=0.5] 99.6462 8.19283 75.4266 97.1788
LR_12[cache_size=All,treshold=0.6] 0 0 0 0
LR_12[cache_size=All,treshold=0.7] 0 0 0 0
LR_12[cache_size=All,treshold=0.8] 0 0 0 0
LR_12[cache_size=All,treshold=0.9] 0 0 0 0
LR_13[cache_size=0.001,treshold=0.3] 66.829 66.8063 66.818 66.8192
LR_13[cache_size=0.001,treshold=0.5] 54.9514 54.9081 54.9317 54.9384
LR_13[cache_size=0.001,treshold=0.6] 46.7527 46.7059 46.727 46.7234
LR_13[cache_size=0.001,treshold=0.7] 35.2222 35.1685 35.1912 35.1846
LR_13[cache_size=0.001,treshold=0.8] 17.4747 17.4552 17.4597 17.4559
LR_13[cache_size=0.001,treshold=0.9] 0 0 0 0
LR_13[cache_size=All,treshold=0.3] 99.9777 67.5232 92.8336 99.8203
LR_13[cache_size=All,treshold=0.5] 99.7419 0.606482 74.5379 98.1443
LR_13[cache_size=All,treshold=0.6] 0 0 0 0
LR_13[cache_size=All,treshold=0.7] 0 0 0 0
LR_13[cache_size=All,treshold=0.8] 0 0 0 0
LR_13[cache_size=All,treshold=0.9] 0 0 0 0
LR_14[cache_size=0.001,treshold=0.3] 67.3954 67.371 67.3851 67.3879
LR_14[cache_size=0.001,treshold=0.5] 55.1991 55.1603 55.1758 55.1762
LR_14[cache_size=0.001,treshold=0.6] 46.0919 46.0451 46.0657 46.055
LR_14[cache_size=0.001,treshold=0.7] 33.9776 33.9478 33.9622 33.9624
LR_14[cache_size=0.001,treshold=0.8] 16.4262 16.3975 16.4127 16.4109
LR_14[cache_size=0.001,treshold=0.9] 0.00016319 5.76087e-05 0.000113279 0.000115193
LR_14[cache_size=All,treshold=0.3] 99.9772 67.7593 92.8719 99.8181
LR_14[cache_size=All,treshold=0.5] 99.7279 0.561001 74.0434 98.0193
LR_14[cache_size=All,treshold=0.6] 0 0 0 0
LR_14[cache_size=All,treshold=0.7] 0 0 0 0
LR_14[cache_size=All,treshold=0.8] 0 0 0 0
LR_14[cache_size=All,treshold=0.9] 0 0 0 0
LR_15[cache_size=0.001,treshold=0.3] 66.963 66.9422 66.9539 66.9532
LR_15[cache_size=0.001,treshold=0.5] 54.8713 54.8295 54.8497 54.8541
LR_15[cache_size=0.001,treshold=0.6] 46.5126 46.461 46.4868 46.4836
LR_15[cache_size=0.001,treshold=0.7] 34.9055 34.8619 34.88 34.8766
LR_15[cache_size=0.001,treshold=0.8] 17.5543 17.5221 17.5365 17.5342
LR_15[cache_size=0.001,treshold=0.9] 0 0 0 0
LR_15[cache_size=All,treshold=0.3] 99.9774 67.7008 92.8671 99.8203
LR_15[cache_size=All,treshold=0.5] 99.7205 0.509145 74.0405 97.9911
LR_15[cache_size=All,treshold=0.6] 0 0 0 0
LR_15[cache_size=All,treshold=0.7] 0 0 0 0
LR_15[cache_size=All,treshold=0.8] 0 0 0 0
LR_15[cache_size=All,treshold=0.9] 0 0 0 0
LR_1[cache_size=0.001,treshold=0.5] 55.2612 55.2205 55.2364 55.2312
LR_1[cache_size=All,treshold=0.5] 91.4718 0 49.3399 70.7458
LR_1_log[cache_size=0.001,treshold=0.5] 48.2013 48.1581 48.182 48.1861
LR_1_log[cache_size=All,treshold=0.5] 79.151 11.676 55.465 68.7038
LR_1_mean[cache_size=0.001,treshold=0.5] 52.3269 51.9306 52.1619 52.2138
LR_1_mean[cache_size=All,treshold=0.5] 98.9308 5.82523 66.6283 90.6499
LR_1_robust_scaler[cache_size=0.001,treshold=0.5] 55.258 55.2176 55.2332 55.2285
LR_1_robust_scaler[cache_size=All,treshold=0.5] 93.1924 1.23205 56.8117 81.7051
LR_1_std_scaler[cache_size=0.001,treshold=0.5] 55.2527 55.2119 55.2281 55.2235
LR_1_std_scaler[cache_size=All,treshold=0.5] 93.185 1.2357 56.8121 81.6936
LR_2[cache_size=0.001,treshold=0.5] 45.8362 45.8061 45.8199 45.8217
LR_2[cache_size=All,treshold=0.5] 81.3849 0 32.0491 19.3932
LR_2_log[cache_size=0.001,treshold=0.5] 58.5159 58.4782 58.4942 58.4946
LR_2_log[cache_size=All,treshold=0.5] 87.9602 1.13336 55.0923 71.1151
LR_2_mean[cache_size=0.001,treshold=0.5] 57.4971 57.4233 57.464 57.4607
LR_2_mean[cache_size=All,treshold=0.5] 97.165 0 54.9107 84.737
LR_3[cache_size=0.001,treshold=0.5] 48.7178 48.6766 48.6986 48.7033
LR_3[cache_size=All,treshold=0.5] 81.3851 0 32.0492 19.3917
LR_3_log[cache_size=0.001,treshold=0.5] 53.9702 53.9394 53.9527 53.9537
LR_3_log[cache_size=All,treshold=0.5] 87.4379 0.990202 54.2743 69.868
LR_3_mean[cache_size=0.001,treshold=0.5] 52.7859 52.6834 52.7408 52.7562
LR_3_mean[cache_size=All,treshold=0.5] 96.8811 0 54.4539 83.3494
LR_4[cache_size=0.001,treshold=0.5] 54.387 54.3278 54.3485 54.3469
LR_4[cache_size=All,treshold=0.5] 80.8001 11.925 61.1372 70.9355
LR_4_log[cache_size=0.001,treshold=0.5] 49.3725 49.3181 49.337 49.3278
LR_4_log[cache_size=All,treshold=0.5] 58.3114 28.941 50.0822 57.16
LR_4_mean[cache_size=0.001,treshold=0.5] 52.5371 52.1425 52.3659 52.4227
LR_4_mean[cache_size=All,treshold=0.5] 73.5403 18.3429 58.2695 68.0734
LR_4_robust_scaler[cache_size=0.001,treshold=0.5] 54.3895 54.331 54.3521 54.3513
LR_4_robust_scaler[cache_size=All,treshold=0.5] 80.8043 11.9283 61.1401 70.9347
LR_4_std_scaler[cache_size=0.001,treshold=0.5] 54.3769 54.3175 54.3394 54.3388
LR_4_std_scaler[cache_size=All,treshold=0.5] 80.8721 11.9649 61.1833 70.9298
LR_5[cache_size=0.001,treshold=0.5] 54.4612 54.41 54.4331 54.4409
LR_5[cache_size=All,treshold=0.5] 80.8047 11.9376 61.1427 70.9327
LR_5_imba[cache_size=0.001,treshold=0.5] 41.9916 41.9597 41.9765 41.9822
LR_5_imba[cache_size=All,treshold=0.5] 47.6865 0 13.467 0
LR_6[cache_size=0.001,treshold=0.5] 54.4628 54.4108 54.4344 54.4422
LR_6[cache_size=All,treshold=0.5] 80.815 11.9411 61.1499 70.9319
LR_6_imba[cache_size=0.001,treshold=0.5] 42.0083 41.9765 41.9933 41.9986
LR_6_imba[cache_size=All,treshold=0.5] 47.6765 0 13.4543 0
LR_7[cache_size=0.001,treshold=0.3] 69.3839 69.3614 69.3742 69.3748
LR_7[cache_size=0.001,treshold=0.5] 55.0665 55.0078 55.0368 55.0426
LR_7[cache_size=0.001,treshold=0.6] 44.7721 44.7306 44.7498 44.7505
LR_7[cache_size=0.001,treshold=0.7] 30.4216 30.3858 30.4051 30.4041
LR_7[cache_size=0.001,treshold=0.8] 10.7908 10.7669 10.7769 10.7716
LR_7[cache_size=0.001,treshold=0.9] 2.08557 2.06788 2.08039 2.08282
LR_7[cache_size=All,treshold=0.3] 99.2371 66.7883 92.212 99.1317
LR_7[cache_size=All,treshold=0.5] 96.6911 6.58047 74.107 95.2891
LR_7[cache_size=All,treshold=0.6] 0 0 0 0
LR_7[cache_size=All,treshold=0.7] 0 0 0 0
LR_7[cache_size=All,treshold=0.8] 0 0 0 0
LR_7[cache_size=All,treshold=0.9] 0 0 0 0
LR_8[cache_size=0.001,treshold=0.3] 70.2804 70.264 70.2712 70.2708
LR_8[cache_size=0.001,treshold=0.5] 54.6318 54.584 54.6119 54.6171
LR_8[cache_size=0.001,treshold=0.6] 43.4294 43.3924 43.414 43.4207
LR_8[cache_size=0.001,treshold=0.7] 29.4672 29.433 29.454 29.4546
LR_8[cache_size=0.001,treshold=0.8] 11.7258 11.7036 11.7137 11.7074
LR_8[cache_size=0.001,treshold=0.9] 1.95965 1.93887 1.95192 1.95418
LR_8[cache_size=All,treshold=0.3] 99.2281 67.5488 92.3776 99.1269
LR_8[cache_size=All,treshold=0.5] 96.5693 5.51532 72.8871 94.775
LR_8[cache_size=All,treshold=0.6] 0 0 0 0
LR_8[cache_size=All,treshold=0.7] 0 0 0 0
LR_8[cache_size=All,treshold=0.8] 0 0 0 0
LR_8[cache_size=All,treshold=0.9] 0 0 0 0
LR_9[cache_size=0.001,treshold=0.3] 69.4067 69.3914 69.4012 69.4034
LR_9[cache_size=0.001,treshold=0.5] 54.7783 54.719 54.7512 54.7566
LR_9[cache_size=0.001,treshold=0.6] 44.4575 44.4136 44.435 44.4393
LR_9[cache_size=0.001,treshold=0.7] 30.3545 30.3241 30.3427 30.3407
LR_9[cache_size=0.001,treshold=0.8] 11.4069 11.3961 11.401 11.399
LR_9[cache_size=0.001,treshold=0.9] 2.0927 2.07646 2.08734 2.08817
LR_9[cache_size=All,treshold=0.3] 99.2329 67.4621 92.3623 99.1293
LR_9[cache_size=All,treshold=0.5] 96.602 4.77935 72.7267 94.7841
LR_9[cache_size=All,treshold=0.6] 0 0 0 0
LR_9[cache_size=All,treshold=0.7] 0 0 0 0
LR_9[cache_size=All,treshold=0.8] 0 0 0 0
LR_9[cache_size=All,treshold=0.9] 0 0 0 0
LR_10[cache_size=0.01,treshold=0.3] 67.8627 67.8449 67.8553 67.8553
LR_10[cache_size=0.01,treshold=0.5] 55.4986 55.4549 55.4816 55.4801
LR_10[cache_size=0.01,treshold=0.6] 46.7602 46.7276 46.7433 46.7427
LR_10[cache_size=0.01,treshold=0.7] 34.4885 34.4587 34.4708 34.4681
LR_10[cache_size=0.01,treshold=0.8] 16.5256 16.4739 16.5061 16.5117
LR_10[cache_size=0.01,treshold=0.9] 0 0 0 0
LR_11[cache_size=0.01,treshold=0.3] 67.8952 67.8824 67.8897 67.8909
LR_11[cache_size=0.01,treshold=0.5] 55.3983 55.3555 55.3813 55.3809
LR_11[cache_size=0.01,treshold=0.6] 46.5881 46.5561 46.5707 46.57
LR_11[cache_size=0.01,treshold=0.7] 34.3995 34.3685 34.3824 34.3775
LR_11[cache_size=0.01,treshold=0.8] 16.5874 16.5366 16.5689 16.5762
LR_11[cache_size=0.01,treshold=0.9] 0 0 0 0
LR_12[cache_size=0.01,treshold=0.3] 67.838 67.8195 67.8298 67.8316
LR_12[cache_size=0.01,treshold=0.5] 55.5696 55.5221 55.549 55.5451
LR_12[cache_size=0.01,treshold=0.6] 46.8503 46.8153 46.8322 46.8322
LR_12[cache_size=0.01,treshold=0.7] 34.5398 34.5128 34.5247 34.5231
LR_12[cache_size=0.01,treshold=0.8] 16.4081 16.3551 16.386 16.3901
LR_12[cache_size=0.01,treshold=0.9] 0 0 0 0
LR_13[cache_size=0.01,treshold=0.3] 74.0248 74.0001 74.0165 74.0177
LR_13[cache_size=0.01,treshold=0.5] 58.5534 58.5279 58.5398 58.5403
LR_13[cache_size=0.01,treshold=0.6] 46.9892 46.9501 46.9768 46.9833
LR_13[cache_size=0.01,treshold=0.7] 32.4144 32.3608 32.3965 32.4004
LR_13[cache_size=0.01,treshold=0.8] 7.82267 7.80438 7.81444 7.81283
LR_13[cache_size=0.01,treshold=0.9] 0 0 0 0
LR_14[cache_size=0.01,treshold=0.3] 73.989 73.964 73.9815 73.9852
LR_14[cache_size=0.01,treshold=0.5] 58.2852 58.2595 58.2743 58.2714
LR_14[cache_size=0.01,treshold=0.6] 46.806 46.7635 46.7913 46.7958
LR_14[cache_size=0.01,treshold=0.7] 32.3267 32.2724 32.307 32.3086
LR_14[cache_size=0.01,treshold=0.8] 8.32671 8.30491 8.31782 8.31693
LR_14[cache_size=0.01,treshold=0.9] 0 0 0 0
LR_15[cache_size=0.01,treshold=0.3] 74.0456 74.0226 74.0384 74.0432
LR_15[cache_size=0.01,treshold=0.5] 58.6472 58.6192 58.6329 58.633
LR_15[cache_size=0.01,treshold=0.6] 47.0552 47.0175 47.043 47.0512
LR_15[cache_size=0.01,treshold=0.7] 32.4456 32.3889 32.4258 32.4263
LR_15[cache_size=0.01,treshold=0.8] 7.64665 7.62795 7.63708 7.63426
LR_15[cache_size=0.01,treshold=0.9] 0 0 0 0
LR_1[cache_size=0.01,treshold=0.5] 54.1189 54.0982 54.1064 54.1065
LR_1_log[cache_size=0.01,treshold=0.5] 49.0082 48.9813 48.9945 48.9949
LR_1_mean[cache_size=0.01,treshold=0.5] 53.7278 53.2604 53.5578 53.6823
LR_1_robust_scaler[cache_size=0.01,treshold=0.5] 54.1265 54.103 54.1129 54.1133
LR_1_std_scaler[cache_size=0.01,treshold=0.5] 54.1292 54.1063 54.116 54.1168
LR_2[cache_size=0.01,treshold=0.5] 45.8245 45.7926 45.8099 45.8127
LR_2_log[cache_size=0.01,treshold=0.5] 58.0402 57.9786 58.0084 58.0127
LR_2_mean[cache_size=0.01,treshold=0.5] 57.8796 57.8278 57.8543 57.8521
LR_3[cache_size=0.01,treshold=0.5] 45.633 45.6 45.6171 45.6197
LR_3_log[cache_size=0.01,treshold=0.5] 57.4241 57.3728 57.3967 57.4004
LR_3_mean[cache_size=0.01,treshold=0.5] 57.2164 57.1678 57.1952 57.1968
LR_4[cache_size=0.01,treshold=0.5] 53.9201 53.9058 53.912 53.9113
LR_4_log[cache_size=0.01,treshold=0.5] 50.1762 50.1481 50.163 50.164
LR_4_mean[cache_size=0.01,treshold=0.5] 54.3545 53.9765 54.2102 54.3158
LR_4_robust_scaler[cache_size=0.01,treshold=0.5] 53.9261 53.9111 53.9172 53.9157
LR_4_std_scaler[cache_size=0.01,treshold=0.5] 53.9213 53.9068 53.9127 53.9113
LR_5[cache_size=0.01,treshold=0.5] 53.9286 53.9138 53.9211 53.9193
LR_5_imba[cache_size=0.01,treshold=0.5] 41.3928 41.358 41.3713 41.3662
LR_6[cache_size=0.01,treshold=0.5] 53.9196 53.9049 53.9117 53.9102
LR_6_imba[cache_size=0.01,treshold=0.5] 41.4103 41.3748 41.3892 41.384
LR_7[cache_size=0.01,treshold=0.3] 68.8959 68.8611 68.8799 68.8839
LR_7[cache_size=0.01,treshold=0.5] 54.3901 54.3474 54.3671 54.3647
LR_7[cache_size=0.01,treshold=0.6] 44.4902 44.4516 44.4727 44.4755
LR_7[cache_size=0.01,treshold=0.7] 31.1963 31.1736 31.1881 31.1898
LR_7[cache_size=0.01,treshold=0.8] 13.1008 13.0903 13.095 13.0926
LR_7[cache_size=0.01,treshold=0.9] 1.55324 1.54283 1.54949 1.55177
LR_8[cache_size=0.01,treshold=0.3] 68.8875 68.8517 68.8704 68.872
LR_8[cache_size=0.01,treshold=0.5] 54.3587 54.3158 54.3351 54.3318
LR_8[cache_size=0.01,treshold=0.6] 44.4638 44.4248 44.4453 44.4475
LR_8[cache_size=0.01,treshold=0.7] 31.1961 31.1752 31.1887 31.189
LR_8[cache_size=0.01,treshold=0.8] 13.146 13.1354 13.141 13.1401
LR_8[cache_size=0.01,treshold=0.9] 1.54754 1.53787 1.54443 1.54725
LR_9[cache_size=0.01,treshold=0.3] 69.0317 68.9917 69.0127 69.0161
LR_9[cache_size=0.01,treshold=0.5] 54.6412 54.5996 54.6155 54.6095
LR_9[cache_size=0.01,treshold=0.6] 44.6607 44.6221 44.6421 44.6479
LR_9[cache_size=0.01,treshold=0.7] 31.1167 31.0942 31.1081 31.1102
LR_9[cache_size=0.01,treshold=0.8] 12.6351 12.6235 12.6298 12.6296
LR_9[cache_size=0.01,treshold=0.9] 1.54677 1.5375 1.54328 1.54577
LR_10[cache_size=0.1,treshold=0.3] 74.9456 74.9101 74.9288 74.9259
LR_10[cache_size=0.1,treshold=0.5] 62.6098 62.5507 62.584 62.5891
LR_10[cache_size=0.1,treshold=0.6] 51.9497 51.9058 51.9295 51.9351
LR_10[cache_size=0.1,treshold=0.7] 21.6587 21.6172 21.6426 21.6488
LR_10[cache_size=0.1,treshold=0.8] 0.0109785 0.00986522 0.0103329 0.0100063
LR_10[cache_size=0.1,treshold=0.9] 0.00753458 0.00630925 0.00684968 0.00671965
LR_11[cache_size=0.1,treshold=0.3] 74.9828 74.9523 74.9681 74.9644
LR_11[cache_size=0.1,treshold=0.5] 63.0481 62.9884 63.0237 63.0299
LR_11[cache_size=0.1,treshold=0.6] 52.305 52.261 52.2842 52.284
LR_11[cache_size=0.1,treshold=0.7] 20.7968 20.7638 20.7788 20.7795
LR_11[cache_size=0.1,treshold=0.8] 0 0 0 0
LR_11[cache_size=0.1,treshold=0.9] 0 0 0 0
LR_12[cache_size=0.1,treshold=0.3] 74.9838 74.9532 74.969 74.9648
LR_12[cache_size=0.1,treshold=0.5] 63.3339 63.2755 63.3056 63.3126
LR_12[cache_size=0.1,treshold=0.6] 52.51 52.4658 52.4853 52.4828
LR_12[cache_size=0.1,treshold=0.7] 20.444 20.4068 20.4248 20.4304
LR_12[cache_size=0.1,treshold=0.8] 0 0 0 0
LR_12[cache_size=0.1,treshold=0.9] 0 0 0 0
LR_13[cache_size=0.1,treshold=0.3] 75.0436 75.0123 75.025 75.0142
LR_13[cache_size=0.1,treshold=0.5] 56.5272 56.4923 56.5077 56.5129
LR_13[cache_size=0.1,treshold=0.6] 43.2788 43.2088 43.2477 43.2431
LR_13[cache_size=0.1,treshold=0.7] 23.8681 23.839 23.8544 23.8565
LR_13[cache_size=0.1,treshold=0.8] 0.00270622 0.00188888 0.00226051 0.00223988
LR_13[cache_size=0.1,treshold=0.9] 0.000214161 0.000136341 0.00018886 0.000194675
LR_14[cache_size=0.1,treshold=0.3] 75.2085 75.1766 75.1906 75.1814
LR_14[cache_size=0.1,treshold=0.5] 56.5957 56.564 56.579 56.5836
LR_14[cache_size=0.1,treshold=0.6] 43.2875 43.2164 43.2557 43.2506
LR_14[cache_size=0.1,treshold=0.7] 23.6399 23.6103 23.6243 23.6245
LR_14[cache_size=0.1,treshold=0.8] 0 0 0 0
LR_14[cache_size=0.1,treshold=0.9] 0 0 0 0
LR_15[cache_size=0.1,treshold=0.3] 75.2561 75.2234 75.2383 75.2306
LR_15[cache_size=0.1,treshold=0.5] 56.6606 56.6298 56.6441 56.6489
LR_15[cache_size=0.1,treshold=0.6] 43.2663 43.1944 43.2335 43.23
LR_15[cache_size=0.1,treshold=0.7] 23.5367 23.5091 23.5222 23.524
LR_15[cache_size=0.1,treshold=0.8] 0 0 0 0
LR_15[cache_size=0.1,treshold=0.9] 0 0 0 0
LR_1[cache_size=0.1,treshold=0.5] 50.7059 50.6403 50.6794 50.6869
LR_1_log[cache_size=0.1,treshold=0.5] 50.5629 50.5054 50.5391 50.5473
LR_1_mean[cache_size=0.1,treshold=0.5] 56.2494 56.1883 56.2282 56.234
LR_1_robust_scaler[cache_size=0.1,treshold=0.5] 50.7024 50.6359 50.6755 50.6831
LR_1_std_scaler[cache_size=0.1,treshold=0.5] 50.7033 50.6362 50.6762 50.6836
LR_2[cache_size=0.1,treshold=0.5] 45.2205 45.1853 45.2007 45.1962
LR_2_log[cache_size=0.1,treshold=0.5] 56.7584 56.7193 56.7391 56.7443
LR_2_mean[cache_size=0.1,treshold=0.5] 56.6635 56.6314 56.6459 56.6496
LR_3[cache_size=0.1,treshold=0.5] 45.2239 45.1878 45.2031 45.1982
LR_3_log[cache_size=0.1,treshold=0.5] 56.7069 56.6659 56.6872 56.6925
LR_3_mean[cache_size=0.1,treshold=0.5] 56.6137 56.58 56.5949 56.597
LR_4[cache_size=0.1,treshold=0.5] 51.9413 51.8676 51.912 51.9257
LR_4_log[cache_size=0.1,treshold=0.5] 51.481 51.4141 51.4539 51.467
LR_4_mean[cache_size=0.1,treshold=0.5] 57.5785 57.4779 57.538 57.5545
LR_4_robust_scaler[cache_size=0.1,treshold=0.5] 51.9373 51.8626 51.9079 51.9215
LR_4_std_scaler[cache_size=0.1,treshold=0.5] 51.9459 51.8691 51.9152 51.9286
LR_5[cache_size=0.1,treshold=0.5] 51.958 51.8825 51.9285 51.9419
LR_5_imba[cache_size=0.1,treshold=0.5] 38.4618 38.3938 38.423 38.4113
LR_6[cache_size=0.1,treshold=0.5] 51.9456 51.8704 51.9155 51.9291
LR_6_imba[cache_size=0.1,treshold=0.5] 38.469 38.3999 38.4297 38.418
LR_7[cache_size=0.1,treshold=0.3] 68.923 68.8922 68.9102 68.9191
LR_7[cache_size=0.1,treshold=0.5] 54.548 54.4748 54.5193 54.5296
LR_7[cache_size=0.1,treshold=0.6] 44.6964 44.6353 44.6727 44.6912
LR_7[cache_size=0.1,treshold=0.7] 31.3573 31.2821 31.311 31.2999
LR_7[cache_size=0.1,treshold=0.8] 12.5234 12.4943 12.5057 12.5024
LR_7[cache_size=0.1,treshold=0.9] 0.00770981 0.00649424 0.00700548 0.00685185
LR_8[cache_size=0.1,treshold=0.3] 69.0532 69.0238 69.041 69.0467
LR_8[cache_size=0.1,treshold=0.5] 54.6622 54.5917 54.6359 54.6467
LR_8[cache_size=0.1,treshold=0.6] 44.7504 44.6907 44.7268 44.7461
LR_8[cache_size=0.1,treshold=0.7] 31.2825 31.2146 31.2384 31.2272
LR_8[cache_size=0.1,treshold=0.8] 12.2738 12.2486 12.2584 12.258
LR_8[cache_size=0.1,treshold=0.9] 0.00202563 0.001441 0.00170174 0.00156697
LR_9[cache_size=0.1,treshold=0.3] 69.0755 69.0468 69.0634 69.0689
LR_9[cache_size=0.1,treshold=0.5] 54.6932 54.6235 54.6677 54.6782
LR_9[cache_size=0.1,treshold=0.6] 44.7754 44.7194 44.7534 44.7722
LR_9[cache_size=0.1,treshold=0.7] 31.2972 31.2263 31.2513 31.24
LR_9[cache_size=0.1,treshold=0.8] 12.2528 12.2285 12.2387 12.2386
LR_9[cache_size=0.1,treshold=0.9] 0.00206459 0.00149942 0.00175042 0.00162537
LR_10[cache_size=0.2,treshold=0.3] 74.4583 74.4314 74.4462 74.449
LR_10[cache_size=0.2,treshold=0.5] 56.1385 56.1264 56.1318 56.1288
LR_10[cache_size=0.2,treshold=0.6] 42.9998 42.9608 42.9763 42.9755
LR_10[cache_size=0.2,treshold=0.7] 23.5909 23.5487 23.5701 23.5765
LR_10[cache_size=0.2,treshold=0.8] 0.00441873 0.00392863 0.00420447 0.00421954
LR_10[cache_size=0.2,treshold=0.9] 0.00297914 0.00256911 0.00273501 0.00270989
LR_11[cache_size=0.2,treshold=0.3] 74.5336 74.51 74.5207 74.5214
LR_11[cache_size=0.2,treshold=0.5] 56.1767 56.1642 56.1701 56.1677
LR_11[cache_size=0.2,treshold=0.6] 42.9901 42.9512 42.9672 42.9654
LR_11[cache_size=0.2,treshold=0.7] 23.4647 23.4244 23.4461 23.4549
LR_11[cache_size=0.2,treshold=0.8] 0 0 0 0
LR_11[cache_size=0.2,treshold=0.9] 0 0 0 0
LR_12[cache_size=0.2,treshold=0.3] 74.5689 74.5431 74.5558 74.5573
LR_12[cache_size=0.2,treshold=0.5] 56.193 56.1809 56.1865 56.1838
LR_12[cache_size=0.2,treshold=0.6] 42.9788 42.9402 42.9555 42.9518
LR_12[cache_size=0.2,treshold=0.7] 23.4095 23.3668 23.3908 23.4003
LR_12[cache_size=0.2,treshold=0.8] 0 0 0 0
LR_12[cache_size=0.2,treshold=0.9] 0 0 0 0
LR_13[cache_size=0.2,treshold=0.3] 74.5525 74.5278 74.5402 74.5423
LR_13[cache_size=0.2,treshold=0.5] 56.1839 56.1722 56.1774 56.174
LR_13[cache_size=0.2,treshold=0.6] 42.9849 42.9466 42.9625 42.9601
LR_13[cache_size=0.2,treshold=0.7] 23.4421 23.4012 23.4231 23.4328
LR_13[cache_size=0.2,treshold=0.8] 7.9977e-05 9.99959e-06 3.19886e-05 1.99978e-05
LR_13[cache_size=0.2,treshold=0.9] 9.99653e-06 0 1.99931e-06 0
LR_14[cache_size=0.2,treshold=0.3] 74.5338 74.5102 74.521 74.5217
LR_14[cache_size=0.2,treshold=0.5] 56.1768 56.1643 56.1703 56.1678
LR_14[cache_size=0.2,treshold=0.6] 42.9902 42.9513 42.9673 42.9654
LR_14[cache_size=0.2,treshold=0.7] 23.4647 23.4244 23.4461 23.455
LR_14[cache_size=0.2,treshold=0.8] 9.99959e-06 0 5.99901e-06 9.99653e-06
LR_14[cache_size=0.2,treshold=0.9] 0 0 0 0
LR_15[cache_size=0.2,treshold=0.3] 74.5683 74.5423 74.555 74.5567
LR_15[cache_size=0.2,treshold=0.5] 56.1948 56.1823 56.1884 56.1859
LR_15[cache_size=0.2,treshold=0.6] 42.9813 42.9424 42.9582 42.9546
LR_15[cache_size=0.2,treshold=0.7] 23.4121 23.3698 23.3935 23.4032
LR_15[cache_size=0.2,treshold=0.8] 0.000109968 9.98949e-06 6.59849e-05 6.99757e-05
LR_15[cache_size=0.2,treshold=0.9] 3.99885e-05 0 1.99951e-05 1.99978e-05
LR_1[cache_size=0.2,treshold=0.5] 48.9285 48.8866 48.9053 48.9058
LR_1_log[cache_size=0.2,treshold=0.5] 51.6398 51.583 51.6184 51.6212
LR_1_mean[cache_size=0.2,treshold=0.5] 56.1303 56.1189 56.1236 56.1234
LR_1_robust_scaler[cache_size=0.2,treshold=0.5] 48.9361 48.8951 48.9127 48.9129
LR_1_std_scaler[cache_size=0.2,treshold=0.5] 48.915 48.8762 48.8935 48.8944
LR_2[cache_size=0.2,treshold=0.5] 45.075 45.026 45.0504 45.0499
LR_2_log[cache_size=0.2,treshold=0.5] 56.0714 56.0589 56.0675 56.07
LR_2_mean[cache_size=0.2,treshold=0.5] 56.1735 56.1604 56.1658 56.1638
LR_3[cache_size=0.2,treshold=0.5] 45.0802 45.0312 45.0553 45.0549
LR_3_log[cache_size=0.2,treshold=0.5] 56.0529 56.0376 56.0471 56.0493
LR_3_mean[cache_size=0.2,treshold=0.5] 56.151 56.1385 56.1437 56.142
LR_4[cache_size=0.2,treshold=0.5] 49.6087 49.5556 49.5781 49.5801
LR_4_log[cache_size=0.2,treshold=0.5] 52.3529 52.3202 52.3315 52.3282
LR_4_mean[cache_size=0.2,treshold=0.5] 57.2244 57.196 57.2055 57.2024
LR_4_robust_scaler[cache_size=0.2,treshold=0.5] 49.6061 49.5576 49.5778 49.579
LR_4_std_scaler[cache_size=0.2,treshold=0.5] 49.5937 49.5449 49.5646 49.5642
LR_5[cache_size=0.2,treshold=0.5] 49.5957 49.5459 49.5672 49.5682
LR_5_imba[cache_size=0.2,treshold=0.5] 35.6025 35.5926 35.5958 35.5945
LR_6[cache_size=0.2,treshold=0.5] 49.606 49.5576 49.5778 49.5788
LR_6_imba[cache_size=0.2,treshold=0.5] 35.6277 35.62 35.6239 35.6227
LR_7[cache_size=0.2,treshold=0.3] 69.5891 69.5646 69.5797 69.5878
LR_7[cache_size=0.2,treshold=0.5] 54.8336 54.7859 54.8112 54.8093
LR_7[cache_size=0.2,treshold=0.6] 44.5147 44.4682 44.4938 44.4933
LR_7[cache_size=0.2,treshold=0.7] 30.2947 30.2775 30.2891 30.2932
LR_7[cache_size=0.2,treshold=0.8] 10.4374 10.3694 10.4005 10.3963
LR_7[cache_size=0.2,treshold=0.9] 0.00777916 0.00636331 0.00716753 0.00743969
LR_8[cache_size=0.2,treshold=0.3] 69.7063 69.6769 69.6931 69.6982
LR_8[cache_size=0.2,treshold=0.5] 54.9255 54.8923 54.9038 54.9009
LR_8[cache_size=0.2,treshold=0.6] 44.5447 44.5027 44.5249 44.5246
LR_8[cache_size=0.2,treshold=0.7] 30.2211 30.2054 30.2151 30.2175
LR_8[cache_size=0.2,treshold=0.8] 10.2106 10.1508 10.1785 10.1771
LR_8[cache_size=0.2,treshold=0.9] 0 0 0 0
LR_9[cache_size=0.2,treshold=0.3] 69.7393 69.7092 69.7248 69.7293
LR_9[cache_size=0.2,treshold=0.5] 54.952 54.9204 54.9308 54.9263
LR_9[cache_size=0.2,treshold=0.6] 44.5602 44.5161 44.5394 44.5399
LR_9[cache_size=0.2,treshold=0.7] 30.2082 30.1953 30.2029 30.2049
LR_9[cache_size=0.2,treshold=0.8] 10.158 10.0991 10.1267 10.1253
LR_9[cache_size=0.2,treshold=0.9] 0 0 0 0
LR_10[cache_size=0.4,treshold=0.3] 74.7041 74.6631 74.6901 74.6924
LR_10[cache_size=0.4,treshold=0.5] 56.2596 56.2169 56.2384 56.2369
LR_10[cache_size=0.4,treshold=0.6] 42.7017 42.5409 42.6008 42.5851
LR_10[cache_size=0.4,treshold=0.7] 21.6635 21.515 21.5831 21.5802
LR_10[cache_size=0.4,treshold=0.8] 0 0 0 0
LR_10[cache_size=0.4,treshold=0.9] 0 0 0 0
LR_11[cache_size=0.4,treshold=0.3] 74.6961 74.6552 74.6818 74.6839
LR_11[cache_size=0.4,treshold=0.5] 56.2589 56.2152 56.2375 56.2363
LR_11[cache_size=0.4,treshold=0.6] 42.7111 42.5457 42.6074 42.5912
LR_11[cache_size=0.4,treshold=0.7] 21.6802 21.5303 21.5984 21.5947
LR_11[cache_size=0.4,treshold=0.8] 0 0 0 0
LR_11[cache_size=0.4,treshold=0.9] 0 0 0 0
LR_12[cache_size=0.4,treshold=0.3] 74.7116 74.6719 74.6982 74.7006
LR_12[cache_size=0.4,treshold=0.5] 56.2644 56.2215 56.2429 56.2407
LR_12[cache_size=0.4,treshold=0.6] 42.701 42.5402 42.5999 42.584
LR_12[cache_size=0.4,treshold=0.7] 21.6484 21.4979 21.5674 21.5647
LR_12[cache_size=0.4,treshold=0.8] 6.49243e-05 0 1.94808e-05 1.08217e-05
LR_12[cache_size=0.4,treshold=0.9] 0 0 0 0
LR_13[cache_size=0.4,treshold=0.3] 97.245 97.2333 97.2396 97.239
LR_13[cache_size=0.4,treshold=0.5] 85.5184 85.4949 85.5064 85.5049
LR_13[cache_size=0.4,treshold=0.6] 0 0 0 0
LR_13[cache_size=0.4,treshold=0.7] 0 0 0 0
LR_13[cache_size=0.4,treshold=0.8] 0 0 0 0
LR_13[cache_size=0.4,treshold=0.9] 0 0 0 0
LR_14[cache_size=0.4,treshold=0.3] 97.245 97.2333 97.2396 97.239
LR_14[cache_size=0.4,treshold=0.5] 85.5184 85.4949 85.5064 85.5049
LR_14[cache_size=0.4,treshold=0.6] 0 0 0 0
LR_14[cache_size=0.4,treshold=0.7] 0 0 0 0
LR_14[cache_size=0.4,treshold=0.8] 0 0 0 0
LR_14[cache_size=0.4,treshold=0.9] 0 0 0 0
LR_15[cache_size=0.4,treshold=0.3] 97.245 97.2333 97.2396 97.239
LR_15[cache_size=0.4,treshold=0.5] 85.5184 85.4949 85.5064 85.5049
LR_15[cache_size=0.4,treshold=0.6] 0 0 0 0
LR_15[cache_size=0.4,treshold=0.7] 0 0 0 0
LR_15[cache_size=0.4,treshold=0.8] 0 0 0 0
LR_15[cache_size=0.4,treshold=0.9] 0 0 0 0
LR_1[cache_size=0.4,treshold=0.5] 47.679 47.6248 47.6513 47.6471
LR_1_log[cache_size=0.4,treshold=0.5] 53.6004 53.5133 53.5535 53.5565
LR_1_mean[cache_size=0.4,treshold=0.5] 56.1774 56.1442 56.1597 56.1563
LR_1_robust_scaler[cache_size=0.4,treshold=0.5] 47.6793 47.6217 47.6501 47.6458
LR_1_std_scaler[cache_size=0.4,treshold=0.5] 47.6618 47.6099 47.6353 47.63
LR_2[cache_size=0.4,treshold=0.5] 52.7435 52.6838 52.7129 52.7113
LR_2_log[cache_size=0.4,treshold=0.5] 55.5793 55.5449 55.5696 55.5757
LR_2_mean[cache_size=0.4,treshold=0.5] 56.1826 56.1433 56.1618 56.157
LR_3[cache_size=0.4,treshold=0.5] 45.4591 45.3881 45.4196 45.421
LR_3_log[cache_size=0.4,treshold=0.5] 55.5817 55.5481 55.5725 55.579
LR_3_mean[cache_size=0.4,treshold=0.5] 56.1745 56.1366 56.1545 56.1498
LR_4[cache_size=0.4,treshold=0.5] 49.0559 48.9897 49.0205 49.0172
LR_4_log[cache_size=0.4,treshold=0.5] 53.8212 53.7175 53.7639 53.7702
LR_4_mean[cache_size=0.4,treshold=0.5] 57.2367 57.1551 57.1941 57.1944
LR_4_robust_scaler[cache_size=0.4,treshold=0.5] 47.7227 47.6629 47.6895 47.6836
LR_4_std_scaler[cache_size=0.4,treshold=0.5] 47.6924 47.6328 47.6584 47.6522
LR_5[cache_size=0.4,treshold=0.5] 47.7061 47.6488 47.6741 47.6681
LR_5_imba[cache_size=0.4,treshold=0.5] 30.4881 30.3955 30.4244 30.4108
LR_6[cache_size=0.4,treshold=0.5] 47.7233 47.6627 47.6897 47.6849
LR_6_imba[cache_size=0.4,treshold=0.5] 30.6833 30.5817 30.6156 30.6029
LR_7[cache_size=0.4,treshold=0.3] 71.2408 71.1857 71.2103 71.2129
LR_7[cache_size=0.4,treshold=0.5] 56.0155 55.9463 55.9762 55.9707
LR_7[cache_size=0.4,treshold=0.6] 44.7207 44.6566 44.6876 44.6848
LR_7[cache_size=0.4,treshold=0.7] 28.1993 28.0878 28.1311 28.1207
LR_7[cache_size=0.4,treshold=0.8] 6.74312 6.65097 6.69731 6.71616
LR_7[cache_size=0.4,treshold=0.9] 0 0 0 0
LR_8[cache_size=0.4,treshold=0.3] 71.2314 71.1766 71.201 71.2032
LR_8[cache_size=0.4,treshold=0.5] 56.0076 55.939 55.9691 55.964
LR_8[cache_size=0.4,treshold=0.6] 44.7167 44.653 44.6843 44.6826
LR_8[cache_size=0.4,treshold=0.7] 28.2041 28.0924 28.1352 28.1246
LR_8[cache_size=0.4,treshold=0.8] 6.75206 6.66728 6.7124 6.72824
LR_8[cache_size=0.4,treshold=0.9] 0 0 0 0
LR_9[cache_size=0.4,treshold=0.3] 71.2425 71.1869 71.2115 71.2141
LR_9[cache_size=0.4,treshold=0.5] 56.0155 55.947 55.9765 55.9711
LR_9[cache_size=0.4,treshold=0.6] 44.718 44.654 44.6854 44.6834
LR_9[cache_size=0.4,treshold=0.7] 28.1928 28.081 28.1242 28.114
LR_9[cache_size=0.4,treshold=0.8] 6.73749 6.63896 6.68599 6.69981
LR_9[cache_size=0.4,treshold=0.9] 0 0 0 0
Offline Clock 1st iteration 0 0 0 0
Offline Clock 2nd iteration 43.4673 41.4753 42.6567 43.0175
Zipf Optimal Distribution 17.0869 6.59027 12.1282 13.1121

Miss Ratio Reduced (%)

Model Max Min Avg Mdn
LR_10[cache_size=0.001,treshold=0.3] -3.21687 -3.2229 -3.2198 -3.21918
LR_10[cache_size=0.001,treshold=0.5] -2.17022 -2.17461 -2.17256 -2.17253
LR_10[cache_size=0.001,treshold=0.6] -1.58558 -1.591 -1.58861 -1.58953
LR_10[cache_size=0.001,treshold=0.7] -0.931972 -0.93897 -0.936225 -0.936453
LR_10[cache_size=0.001,treshold=0.8] -0.267959 -0.269508 -0.268521 -0.268143
LR_10[cache_size=0.001,treshold=0.9] 0.000682298 0.000341196 0.000511724 0.000511706
LR_10[cache_size=All,treshold=0.3] -3.25695 -26.6916 -15.5345 -17.181
LR_10[cache_size=All,treshold=0.5] -0.0279728 -26.4195 -13.7829 -16.2085
LR_10[cache_size=All,treshold=0.6] 0 0 0 0
LR_10[cache_size=All,treshold=0.7] 0 0 0 0
LR_10[cache_size=All,treshold=0.8] 0 0 0 0
LR_10[cache_size=All,treshold=0.9] 0 0 0 0
LR_11[cache_size=0.001,treshold=0.3] -3.2312 -3.23603 -3.23314 -3.23256
LR_11[cache_size=0.001,treshold=0.5] -2.15299 -2.15687 -2.15499 -2.15514
LR_11[cache_size=0.001,treshold=0.6] -1.5559 -1.56165 -1.5592 -1.56036
LR_11[cache_size=0.001,treshold=0.7] -0.908775 -0.914916 -0.912481 -0.912402
LR_11[cache_size=0.001,treshold=0.8] -0.270227 -0.272228 -0.271284 -0.27137
LR_11[cache_size=0.001,treshold=0.9] 0.00818756 0.0075049 0.00764181 0.00750528
LR_11[cache_size=All,treshold=0.3] -3.25661 -26.6916 -15.5343 -17.181
LR_11[cache_size=All,treshold=0.5] -0.0286565 -26.4207 -13.787 -16.2117
LR_11[cache_size=All,treshold=0.6] 0 0 0 0
LR_11[cache_size=All,treshold=0.7] 0 0 0 0
LR_11[cache_size=All,treshold=0.8] 0 0 0 0
LR_11[cache_size=All,treshold=0.9] 0 0 0 0
LR_12[cache_size=0.001,treshold=0.3] -3.21499 -3.22051 -3.21748 -3.21696
LR_12[cache_size=0.001,treshold=0.5] -2.16186 -2.16557 -2.16376 -2.16383
LR_12[cache_size=0.001,treshold=0.6] -1.57944 -1.58502 -1.58291 -1.58407
LR_12[cache_size=0.001,treshold=0.7] -0.936577 -0.943235 -0.940626 -0.940546
LR_12[cache_size=0.001,treshold=0.8] -0.286906 -0.288953 -0.287694 -0.287287
LR_12[cache_size=0.001,treshold=0.9] 0.00579952 0.00511724 0.00539021 0.00528853
LR_12[cache_size=All,treshold=0.3] -3.25593 -26.6927 -15.5354 -17.182
LR_12[cache_size=All,treshold=0.5] -0.0267802 -26.4184 -13.7803 -16.2071
LR_12[cache_size=All,treshold=0.6] 0 0 0 0
LR_12[cache_size=All,treshold=0.7] 0 0 0 0
LR_12[cache_size=All,treshold=0.8] 0 0 0 0
LR_12[cache_size=All,treshold=0.9] 0 0 0 0
LR_13[cache_size=0.001,treshold=0.3] -3.52662 -3.53197 -3.52902 -3.5285
LR_13[cache_size=0.001,treshold=0.5] -2.43687 -2.44023 -2.43873 -2.43887
LR_13[cache_size=0.001,treshold=0.6] -1.80015 -1.80615 -1.80436 -1.80502
LR_13[cache_size=0.001,treshold=0.7] -1.06348 -1.06914 -1.06644 -1.06626
LR_13[cache_size=0.001,treshold=0.8] -0.281945 -0.284859 -0.283736 -0.283665
LR_13[cache_size=0.001,treshold=0.9] 0 0 0 0
LR_13[cache_size=All,treshold=0.3] -3.55613 -26.7584 -15.8329 -17.4288
LR_13[cache_size=All,treshold=0.5] 0.00136459 -26.5366 -14.077 -16.7173
LR_13[cache_size=All,treshold=0.6] 0 0 0 0
LR_13[cache_size=All,treshold=0.7] 0 0 0 0
LR_13[cache_size=All,treshold=0.8] 0 0 0 0
LR_13[cache_size=All,treshold=0.9] 0 0 0 0
LR_14[cache_size=0.001,treshold=0.3] -3.54368 -3.54834 -3.54587 -3.54556
LR_14[cache_size=0.001,treshold=0.5] -2.40702 -2.41116 -2.40918 -2.40851
LR_14[cache_size=0.001,treshold=0.6] -1.75393 -1.76057 -1.75864 -1.75914
LR_14[cache_size=0.001,treshold=0.7] -1.02288 -1.02785 -1.02561 -1.02532
LR_14[cache_size=0.001,treshold=0.8] -0.296102 -0.299007 -0.297553 -0.297311
LR_14[cache_size=0.001,treshold=0.9] 0 0 0 0
LR_14[cache_size=All,treshold=0.3] -3.57318 -26.7573 -15.8342 -17.4278
LR_14[cache_size=All,treshold=0.5] 0.00221747 -26.5199 -13.9948 -16.6568
LR_14[cache_size=All,treshold=0.6] 0 0 0 0
LR_14[cache_size=All,treshold=0.7] 0 0 0 0
LR_14[cache_size=All,treshold=0.8] 0 0 0 0
LR_14[cache_size=All,treshold=0.9] 0 0 0 0
LR_15[cache_size=0.001,treshold=0.3] -3.52662 -3.53094 -3.52806 -3.52765
LR_15[cache_size=0.001,treshold=0.5] -2.42476 -2.42822 -2.42645 -2.42625
LR_15[cache_size=0.001,treshold=0.6] -1.78804 -1.79469 -1.79289 -1.79376
LR_15[cache_size=0.001,treshold=0.7] -1.06467 -1.06965 -1.06736 -1.06711
LR_15[cache_size=0.001,treshold=0.8] -0.304289 -0.307546 -0.306048 -0.305541
LR_15[cache_size=0.001,treshold=0.9] 0 0 0 0
LR_15[cache_size=All,treshold=0.3] -3.56738 -26.7584 -15.8349 -17.4288
LR_15[cache_size=All,treshold=0.5] 0.00238804 -26.5154 -13.9961 -16.6522
LR_15[cache_size=All,treshold=0.6] 0 0 0 0
LR_15[cache_size=All,treshold=0.7] 0 0 0 0
LR_15[cache_size=All,treshold=0.8] 0 0 0 0
LR_15[cache_size=All,treshold=0.9] 0 0 0 0
LR_1[cache_size=0.001,treshold=0.5] -2.62791 -2.63318 -2.63076 -2.62991
LR_1[cache_size=All,treshold=0.5] 0 -20.1154 -8.79607 -9.58326
LR_1_log[cache_size=0.001,treshold=0.5] -2.11928 -2.12275 -2.12057 -2.11973
LR_1_log[cache_size=All,treshold=0.5] -0.447075 -12.5428 -6.86296 -8.06001
LR_1_mean[cache_size=0.001,treshold=0.5] -2.29301 -2.32178 -2.30704 -2.30639
LR_1_mean[cache_size=All,treshold=0.5] 0.0274662 -25.9714 -12.57 -15.2456
LR_1_robust_scaler[cache_size=0.001,treshold=0.5] -2.62774 -2.63284 -2.63046 -2.62957
LR_1_robust_scaler[cache_size=All,treshold=0.5] -0.0463939 -20.811 -9.84902 -11.634
LR_1_std_scaler[cache_size=0.001,treshold=0.5] -2.62757 -2.63249 -2.63022 -2.62941
LR_1_std_scaler[cache_size=All,treshold=0.5] -0.0465645 -20.8066 -9.84735 -11.6322
LR_2[cache_size=0.001,treshold=0.5] -2.73792 -2.7427 -2.74014 -2.73993
LR_2[cache_size=All,treshold=0.5] 0 -16.0663 -5.46053 -2.50822
LR_2_log[cache_size=0.001,treshold=0.5] -3.70964 -3.71374 -3.71099 -3.70999
LR_2_log[cache_size=All,treshold=0.5] -0.0835814 -17.2147 -8.97128 -10.5699
LR_2_mean[cache_size=0.001,treshold=0.5] -3.61703 -3.62493 -3.62075 -3.62077
LR_2_mean[cache_size=All,treshold=0.5] 0.000682298 -24.7407 -11.1305 -12.9127
LR_3[cache_size=0.001,treshold=0.5] -2.97642 -2.98051 -2.97771 -2.97721
LR_3[cache_size=All,treshold=0.5] 0 -16.0663 -5.46054 -2.50775
LR_3_log[cache_size=0.001,treshold=0.5] -3.34462 -3.34935 -3.34647 -3.34605
LR_3_log[cache_size=All,treshold=0.5] -0.073174 -16.9461 -8.81658 -10.3848
LR_3_mean[cache_size=0.001,treshold=0.5] -3.25722 -3.26417 -3.26091 -3.26088
LR_3_mean[cache_size=All,treshold=0.5] 0 -24.5512 -10.9792 -12.5683
LR_4[cache_size=0.001,treshold=0.5] -2.48856 -2.49414 -2.49096 -2.49039
LR_4[cache_size=All,treshold=0.5] -0.431361 -18.4488 -10.6101 -13.371
LR_4_log[cache_size=0.001,treshold=0.5] -2.16073 -2.16642 -2.16352 -2.16288
LR_4_log[cache_size=All,treshold=0.5] -1.3095 -8.25893 -5.42126 -6.45446
LR_4_mean[cache_size=0.001,treshold=0.5] -2.28277 -2.31001 -2.29654 -2.29616
LR_4_mean[cache_size=All,treshold=0.5] -0.432555 -20.0229 -10.8456 -11.9511
LR_4_robust_scaler[cache_size=0.001,treshold=0.5] -2.4889 -2.49431 -2.4912 -2.49056
LR_4_robust_scaler[cache_size=All,treshold=0.5] -0.431531 -18.4488 -10.6106 -13.3719
LR_4_std_scaler[cache_size=0.001,treshold=0.5] -2.48804 -2.49346 -2.49031 -2.48971
LR_4_std_scaler[cache_size=All,treshold=0.5] -0.432555 -18.4577 -10.6203 -13.3883
LR_5[cache_size=0.001,treshold=0.5] -2.49721 -2.50079 -2.4986 -2.49823
LR_5[cache_size=All,treshold=0.5] -0.432043 -18.4432 -10.609 -13.3696
LR_5_imba[cache_size=0.001,treshold=0.5] -1.61449 -1.61784 -1.61611 -1.61577
LR_5_imba[cache_size=All,treshold=0.5] 0 -9.23627 -2.34143 0
LR_6[cache_size=0.001,treshold=0.5] -2.49721 -2.50096 -2.4987 -2.49823
LR_6[cache_size=All,treshold=0.5] -0.432043 -18.4488 -10.6121 -13.3738
LR_6_imba[cache_size=0.001,treshold=0.5] -1.61568 -1.61887 -1.61716 -1.61679
LR_6_imba[cache_size=All,treshold=0.5] 0 -9.23259 -2.33873 0
LR_7[cache_size=0.001,treshold=0.3] -3.97043 -3.97459 -3.97269 -3.97284
LR_7[cache_size=0.001,treshold=0.5] -2.68215 -2.6876 -2.68456 -2.68416
LR_7[cache_size=0.001,treshold=0.6] -1.90788 -1.91206 -1.91004 -1.90958
LR_7[cache_size=0.001,treshold=0.7] -1.00809 -1.01181 -1.00964 -1.00895
LR_7[cache_size=0.001,treshold=0.8] -0.118736 -0.120596 -0.119813 -0.119908
LR_7[cache_size=0.001,treshold=0.9] 0.0279742 0.0264382 0.0270193 0.0269494
LR_7[cache_size=All,treshold=0.3] -3.70077 -26.0015 -15.5442 -17.1022
LR_7[cache_size=All,treshold=0.5] -0.231799 -23.5434 -12.9866 -15.4492
LR_7[cache_size=All,treshold=0.6] 0 0 0 0
LR_7[cache_size=All,treshold=0.7] 0 0 0 0
LR_7[cache_size=All,treshold=0.8] 0 0 0 0
LR_7[cache_size=All,treshold=0.9] 0 0 0 0
LR_8[cache_size=0.001,treshold=0.3] -4.01444 -4.01775 -4.01608 -4.01587
LR_8[cache_size=0.001,treshold=0.5] -2.63729 -2.64109 -2.63888 -2.63759
LR_8[cache_size=0.001,treshold=0.6] -1.83384 -1.83685 -1.83502 -1.83453
LR_8[cache_size=0.001,treshold=0.7] -0.98609 -0.990322 -0.98825 -0.988138
LR_8[cache_size=0.001,treshold=0.8] -0.180322 -0.182173 -0.181459 -0.181491
LR_8[cache_size=0.001,treshold=0.9] 0.0262684 0.024732 0.0251088 0.0247333
LR_8[cache_size=All,treshold=0.3] -3.76882 -25.9926 -15.5579 -17.0987
LR_8[cache_size=All,treshold=0.5] -0.198197 -23.433 -12.7635 -15.23
LR_8[cache_size=All,treshold=0.6] 0 0 0 0
LR_8[cache_size=All,treshold=0.7] 0 0 0 0
LR_8[cache_size=All,treshold=0.8] 0 0 0 0
LR_8[cache_size=All,treshold=0.9] 0 0 0 0
LR_9[cache_size=0.001,treshold=0.3] -3.96548 -3.96965 -3.96764 -3.96773
LR_9[cache_size=0.001,treshold=0.5] -2.66181 -2.66627 -2.66403 -2.66403
LR_9[cache_size=0.001,treshold=0.6] -1.89389 -1.89739 -1.89551 -1.89482
LR_9[cache_size=0.001,treshold=0.7] -1.01543 -1.0183 -1.01657 -1.0164
LR_9[cache_size=0.001,treshold=0.8] -0.148591 -0.150271 -0.149664 -0.149935
LR_9[cache_size=0.001,treshold=0.9] 0.0279742 0.0262676 0.0268828 0.0267788
LR_9[cache_size=All,treshold=0.3] -3.75808 -25.997 -15.5573 -17.1001
LR_9[cache_size=All,treshold=0.5] -0.170939 -23.4665 -12.7712 -15.2422
LR_9[cache_size=All,treshold=0.6] 0 0 0 0
LR_9[cache_size=All,treshold=0.7] 0 0 0 0
LR_9[cache_size=All,treshold=0.8] 0 0 0 0
LR_9[cache_size=All,treshold=0.9] 0 0 0 0
LR_10[cache_size=0.01,treshold=0.3] -4.75913 -4.77055 -4.76281 -4.76185
LR_10[cache_size=0.01,treshold=0.5] -3.23771 -3.25293 -3.24348 -3.24177
LR_10[cache_size=0.01,treshold=0.6] -2.36641 -2.37553 -2.36954 -2.36871
LR_10[cache_size=0.01,treshold=0.7] -1.37419 -1.38484 -1.37991 -1.37992
LR_10[cache_size=0.01,treshold=0.8] -0.381734 -0.390036 -0.385496 -0.386683
LR_10[cache_size=0.01,treshold=0.9] 0 0 0 0
LR_11[cache_size=0.01,treshold=0.3] -4.75617 -4.76784 -4.75966 -4.75839
LR_11[cache_size=0.01,treshold=0.5] -3.22981 -3.24479 -3.23563 -3.23412
LR_11[cache_size=0.01,treshold=0.6] -2.36155 -2.37109 -2.36485 -2.36402
LR_11[cache_size=0.01,treshold=0.7] -1.37715 -1.38731 -1.38218 -1.38214
LR_11[cache_size=0.01,treshold=0.8] -0.390178 -0.398917 -0.394232 -0.395566
LR_11[cache_size=0.01,treshold=0.9] 0 0 0 0
LR_12[cache_size=0.01,treshold=0.3] -4.76061 -4.77179 -4.7643 -4.76308
LR_12[cache_size=0.01,treshold=0.5] -3.24116 -3.25713 -3.24723 -3.24537
LR_12[cache_size=0.01,treshold=0.6] -2.36789 -2.37701 -2.37087 -2.3697
LR_12[cache_size=0.01,treshold=0.7] -1.36926 -1.3804 -1.37498 -1.37473
LR_12[cache_size=0.01,treshold=0.8] -0.371864 -0.380908 -0.37607 -0.377059
LR_12[cache_size=0.01,treshold=0.9] 0 0 0 0
LR_13[cache_size=0.01,treshold=0.3] -5.92757 -5.94018 -5.93145 -5.93065
LR_13[cache_size=0.01,treshold=0.5] -4.06459 -4.07702 -4.07 -4.06935
LR_13[cache_size=0.01,treshold=0.6] -2.94703 -2.95874 -2.9525 -2.95253
LR_13[cache_size=0.01,treshold=0.7] -1.70312 -1.71507 -1.71003 -1.71132
LR_13[cache_size=0.01,treshold=0.8] -0.293395 -0.296783 -0.295378 -0.295903
LR_13[cache_size=0.01,treshold=0.9] 0 0 0 0
LR_14[cache_size=0.01,treshold=0.3] -5.91749 -5.93105 -5.92187 -5.92127
LR_14[cache_size=0.01,treshold=0.5] -4.04411 -4.05629 -4.04956 -4.04936
LR_14[cache_size=0.01,treshold=0.6] -2.93617 -2.94812 -2.94184 -2.94217
LR_14[cache_size=0.01,treshold=0.7] -1.7051 -1.71729 -1.71196 -1.7128
LR_14[cache_size=0.01,treshold=0.8] -0.316343 -0.319617 -0.318179 -0.318361
LR_14[cache_size=0.01,treshold=0.9] 0 0 0 0
LR_15[cache_size=0.01,treshold=0.3] -5.93102 -5.94314 -5.93495 -5.93411
LR_15[cache_size=0.01,treshold=0.5] -4.07101 -4.08369 -4.07661 -4.07577
LR_15[cache_size=0.01,treshold=0.6] -2.95048 -2.96194 -2.95541 -2.95525
LR_15[cache_size=0.01,treshold=0.7] -1.70164 -1.71384 -1.70865 -1.71034
LR_15[cache_size=0.01,treshold=0.8] -0.285005 -0.288642 -0.287136 -0.287759
LR_15[cache_size=0.01,treshold=0.9] 0 0 0 0
LR_1[cache_size=0.01,treshold=0.5] -3.30088 -3.31167 -3.30621 -3.30691
LR_1_log[cache_size=0.01,treshold=0.5] -2.8345 -2.84422 -2.83928 -2.83929
LR_1_mean[cache_size=0.01,treshold=0.5] -3.99662 -4.04962 -4.03214 -4.04288
LR_1_robust_scaler[cache_size=0.01,treshold=0.5] -3.30112 -3.31241 -3.3066 -3.30691
LR_1_std_scaler[cache_size=0.01,treshold=0.5] -3.30186 -3.31291 -3.30715 -3.30741
LR_2[cache_size=0.01,treshold=0.5] -3.71197 -3.72137 -3.71727 -3.71805
LR_2_log[cache_size=0.01,treshold=0.5] -4.91393 -4.93098 -4.92148 -4.92004
LR_2_mean[cache_size=0.01,treshold=0.5] -4.94058 -4.95591 -4.94764 -4.94669
LR_3[cache_size=0.01,treshold=0.5] -3.69174 -3.70039 -3.69644 -3.69733
LR_3_log[cache_size=0.01,treshold=0.5] -4.84509 -4.8599 -4.85106 -4.8497
LR_3_mean[cache_size=0.01,treshold=0.5] -4.86976 -4.88507 -4.87697 -4.87562
LR_4[cache_size=0.01,treshold=0.5] -3.15258 -3.16482 -3.15825 -3.15861
LR_4_log[cache_size=0.01,treshold=0.5] -2.88879 -2.90074 -2.89451 -2.89383
LR_4_mean[cache_size=0.01,treshold=0.5] -3.86831 -3.9122 -3.89573 -3.90247
LR_4_robust_scaler[cache_size=0.01,treshold=0.5] -3.15307 -3.16556 -3.15884 -3.15935
LR_4_std_scaler[cache_size=0.01,treshold=0.5] -3.15258 -3.16507 -3.15835 -3.15885
LR_5[cache_size=0.01,treshold=0.5] -3.15356 -3.16581 -3.15929 -3.15935
LR_5_imba[cache_size=0.01,treshold=0.5] -2.01749 -2.02827 -2.02273 -2.02151
LR_6[cache_size=0.01,treshold=0.5] -3.15258 -3.16482 -3.15825 -3.15861
LR_6_imba[cache_size=0.01,treshold=0.5] -2.01872 -2.02975 -2.02407 -2.02295
LR_7[cache_size=0.01,treshold=0.3] -5.19672 -5.21209 -5.20132 -5.19862
LR_7[cache_size=0.01,treshold=0.5] -3.43931 -3.45556 -3.44835 -3.44979
LR_7[cache_size=0.01,treshold=0.6] -2.44611 -2.46043 -2.45378 -2.45409
LR_7[cache_size=0.01,treshold=0.7] -1.33841 -1.34486 -1.34122 -1.34167
LR_7[cache_size=0.01,treshold=0.8] -0.234699 -0.237081 -0.235809 -0.235455
LR_7[cache_size=0.01,treshold=0.9] 0.0399803 0.0380085 0.0390876 0.0389789
LR_8[cache_size=0.01,treshold=0.3] -5.19474 -5.21061 -5.19949 -5.19665
LR_8[cache_size=0.01,treshold=0.5] -3.43684 -3.45334 -3.44608 -3.44757
LR_8[cache_size=0.01,treshold=0.6] -2.44537 -2.45969 -2.45309 -2.4536
LR_8[cache_size=0.01,treshold=0.7] -1.34039 -1.34659 -1.34314 -1.34364
LR_8[cache_size=0.01,treshold=0.8] -0.238368 -0.241028 -0.239461 -0.239157
LR_8[cache_size=0.01,treshold=0.9] 0.0399761 0.0377616 0.0388902 0.0387322
LR_9[cache_size=0.01,treshold=0.3] -5.2172 -5.23283 -5.22145 -5.21849
LR_9[cache_size=0.01,treshold=0.5] -3.45781 -3.47284 -3.46606 -3.46706
LR_9[cache_size=0.01,treshold=0.6] -2.45006 -2.46413 -2.45739 -2.4568
LR_9[cache_size=0.01,treshold=0.7] -1.31917 -1.32585 -1.32227 -1.32267
LR_9[cache_size=0.01,treshold=0.8] -0.20311 -0.20649 -0.205012 -0.205344
LR_9[cache_size=0.01,treshold=0.9] 0.0402229 0.0380085 0.0391863 0.0389789
LR_10[cache_size=0.1,treshold=0.3] -10.324 -10.3411 -10.3285 -10.3258
LR_10[cache_size=0.1,treshold=0.5] -7.98478 -7.99936 -7.9908 -7.98939
LR_10[cache_size=0.1,treshold=0.6] -6.21124 -6.22448 -6.21908 -6.22072
LR_10[cache_size=0.1,treshold=0.7] -2.23289 -2.24142 -2.23817 -2.2393
LR_10[cache_size=0.1,treshold=0.8] -0.00187072 -0.00233845 -0.00224432 -0.00233774
LR_10[cache_size=0.1,treshold=0.9] -0.00140198 -0.00140307 -0.00140271 -0.00140284
LR_11[cache_size=0.1,treshold=0.3] -10.3349 -10.3518 -10.3396 -10.3375
LR_11[cache_size=0.1,treshold=0.5] -8.03948 -8.05502 -8.04579 -8.04363
LR_11[cache_size=0.1,treshold=0.6] -6.25797 -6.27172 -6.26686 -6.27028
LR_11[cache_size=0.1,treshold=0.7] -2.13496 -2.14277 -2.13811 -2.13781
LR_11[cache_size=0.1,treshold=0.8] 0 0 0 0
LR_11[cache_size=0.1,treshold=0.9] 0 0 0 0
LR_12[cache_size=0.1,treshold=0.3] -10.3333 -10.3504 -10.3382 -10.3361
LR_12[cache_size=0.1,treshold=0.5] -8.07314 -8.08776 -8.0787 -8.07728
LR_12[cache_size=0.1,treshold=0.6] -6.28648 -6.30072 -6.29576 -6.29833
LR_12[cache_size=0.1,treshold=0.7] -2.09474 -2.10396 -2.09902 -2.09899
LR_12[cache_size=0.1,treshold=0.8] 0 0 0 0
LR_12[cache_size=0.1,treshold=0.9] 0 0 0 0
LR_13[cache_size=0.1,treshold=0.3] -10.6382 -10.6558 -10.6443 -10.6421
LR_13[cache_size=0.1,treshold=0.5] -7.10103 -7.1159 -7.10896 -7.10953
LR_13[cache_size=0.1,treshold=0.6] -5.02797 -5.04321 -5.0378 -5.03969
LR_13[cache_size=0.1,treshold=0.7] -2.5044 -2.51482 -2.50955 -2.51009
LR_13[cache_size=0.1,treshold=0.8] -0.000467548 -0.000935362 -0.000748096 -0.000934654
LR_13[cache_size=0.1,treshold=0.9] 0 0 0 0
LR_14[cache_size=0.1,treshold=0.3] -10.6733 -10.6895 -10.6785 -10.6762
LR_14[cache_size=0.1,treshold=0.5] -7.11225 -7.12619 -7.12009 -7.12075
LR_14[cache_size=0.1,treshold=0.6] -5.0275 -5.04321 -5.03752 -5.03969
LR_14[cache_size=0.1,treshold=0.7] -2.47683 -2.48583 -2.48103 -2.48063
LR_14[cache_size=0.1,treshold=0.8] 0 0 0 0
LR_14[cache_size=0.1,treshold=0.9] 0 0 0 0
LR_15[cache_size=0.1,treshold=0.3] -10.6836 -10.7003 -10.6889 -10.686
LR_15[cache_size=0.1,treshold=0.5] -7.1244 -7.13741 -7.1315 -7.13244
LR_15[cache_size=0.1,treshold=0.6] -5.0247 -5.03993 -5.03453 -5.03689
LR_15[cache_size=0.1,treshold=0.7] -2.46515 -2.47274 -2.46878 -2.46847
LR_15[cache_size=0.1,treshold=0.8] 0 0 0 0
LR_15[cache_size=0.1,treshold=0.9] 0 0 0 0
LR_1[cache_size=0.1,treshold=0.5] -4.14973 -4.16384 -4.15709 -4.1564
LR_1_log[cache_size=0.1,treshold=0.5] -4.01857 -4.02868 -4.02327 -4.02368
LR_1_mean[cache_size=0.1,treshold=0.5] -7.04833 -7.06071 -7.055 -7.05716
LR_1_robust_scaler[cache_size=0.1,treshold=0.5] -4.14927 -4.16291 -4.15652 -4.15594
LR_1_std_scaler[cache_size=0.1,treshold=0.5] -4.15067 -4.16478 -4.15821 -4.15734
LR_2[cache_size=0.1,treshold=0.5] -5.38921 -5.40088 -5.39681 -5.3983
LR_2_log[cache_size=0.1,treshold=0.5] -7.1286 -7.14162 -7.13589 -7.13758
LR_2_mean[cache_size=0.1,treshold=0.5] -7.12533 -7.13741 -7.13206 -7.13291
LR_3[cache_size=0.1,treshold=0.5] -5.38968 -5.4014 -5.39718 -5.39877
LR_3_log[cache_size=0.1,treshold=0.5] -7.11926 -7.13274 -7.12663 -7.12823
LR_3_mean[cache_size=0.1,treshold=0.5] -7.11599 -7.12853 -7.1228 -7.12355
LR_4[cache_size=0.1,treshold=0.5] -4.00101 -4.01699 -4.00877 -4.00686
LR_4_log[cache_size=0.1,treshold=0.5] -4.04263 -4.05393 -4.04767 -4.04429
LR_4_mean[cache_size=0.1,treshold=0.5] -6.61346 -6.63356 -6.62484 -6.62283
LR_4_robust_scaler[cache_size=0.1,treshold=0.5] -4.00007 -4.01652 -4.0083 -4.00639
LR_4_std_scaler[cache_size=0.1,treshold=0.5] -4.00007 -4.01652 -4.0083 -4.00639
LR_5[cache_size=0.1,treshold=0.5] -4.00382 -4.02073 -4.01223 -4.01013
LR_5_imba[cache_size=0.1,treshold=0.5] -2.51095 -2.52698 -2.519 -2.51868
LR_6[cache_size=0.1,treshold=0.5] -4.00101 -4.01745 -4.00915 -4.00733
LR_6_imba[cache_size=0.1,treshold=0.5] -2.51048 -2.52605 -2.51815 -2.51774
LR_7[cache_size=0.1,treshold=0.3] -7.48836 -7.50548 -7.49648 -7.49505
LR_7[cache_size=0.1,treshold=0.5] -4.77549 -4.79008 -4.78344 -4.78262
LR_7[cache_size=0.1,treshold=0.6] -3.2934 -3.30703 -3.2989 -3.29839
LR_7[cache_size=0.1,treshold=0.7] -1.69359 -1.70679 -1.70037 -1.70187
LR_7[cache_size=0.1,treshold=0.8] -0.231974 -0.239418 -0.235936 -0.236579
LR_7[cache_size=0.1,treshold=0.9] -0.00140198 -0.00140307 -0.00140271 -0.00140284
LR_8[cache_size=0.1,treshold=0.3] -7.5122 -7.52793 -7.51883 -7.51797
LR_8[cache_size=0.1,treshold=0.5] -4.78204 -4.79709 -4.78999 -4.78862
LR_8[cache_size=0.1,treshold=0.6] -3.28639 -3.30095 -3.29227 -3.29092
LR_8[cache_size=0.1,treshold=0.7] -1.67256 -1.68575 -1.67877 -1.6799
LR_8[cache_size=0.1,treshold=0.8] -0.209525 -0.216505 -0.213212 -0.213101
LR_8[cache_size=0.1,treshold=0.9] 0 -0.000467681 -0.000280511 -0.000467327
LR_9[cache_size=0.1,treshold=0.3] -7.51501 -7.53074 -7.52154 -7.52078
LR_9[cache_size=0.1,treshold=0.5] -4.78297 -4.7985 -4.79111 -4.78917
LR_9[cache_size=0.1,treshold=0.6] -3.28546 -3.30049 -3.2917 -3.29045
LR_9[cache_size=0.1,treshold=0.7] -1.66976 -1.68341 -1.67615 -1.67714
LR_9[cache_size=0.1,treshold=0.8] -0.205783 -0.212734 -0.209565 -0.209362
LR_9[cache_size=0.1,treshold=0.9] 0 -0.000467681 -0.000280511 -0.000467327
LR_10[cache_size=0.2,treshold=0.3] -12.2754 -12.2815 -12.2776 -12.2778
LR_10[cache_size=0.2,treshold=0.5] -8.05348 -8.06355 -8.0599 -8.06154
LR_10[cache_size=0.2,treshold=0.6] -5.64452 -5.66228 -5.6566 -5.65871
LR_10[cache_size=0.2,treshold=0.7] -2.75628 -2.76833 -2.7616 -2.76271
LR_10[cache_size=0.2,treshold=0.8] -0.00196289 -0.00196456 -0.00196369 -0.00196354
LR_10[cache_size=0.2,treshold=0.9] -0.00130859 -0.00130971 -0.00130913 -0.00130903
LR_11[cache_size=0.2,treshold=0.3] -12.292 -12.2985 -12.2953 -12.2955
LR_11[cache_size=0.2,treshold=0.5] -8.05938 -8.07009 -8.06579 -8.06743
LR_11[cache_size=0.2,treshold=0.6] -5.64124 -5.65966 -5.65346 -5.65544
LR_11[cache_size=0.2,treshold=0.7] -2.73663 -2.75001 -2.74236 -2.74307
LR_11[cache_size=0.2,treshold=0.8] 0 0 0 0
LR_11[cache_size=0.2,treshold=0.9] 0 0 0 0
LR_12[cache_size=0.2,treshold=0.3] -12.3018 -12.3083 -12.3045 -12.304
LR_12[cache_size=0.2,treshold=0.5] -8.06265 -8.07271 -8.0688 -8.07041
LR_12[cache_size=0.2,treshold=0.6] -5.63928 -5.6577 -5.6515 -5.65335
LR_12[cache_size=0.2,treshold=0.7] -2.72922 -2.74215 -2.73463 -2.73522
LR_12[cache_size=0.2,treshold=0.8] 0 0 0 0
LR_12[cache_size=0.2,treshold=0.9] 0 0 0 0
LR_13[cache_size=0.2,treshold=0.3] -12.2978 -12.3044 -12.3007 -12.3008
LR_13[cache_size=0.2,treshold=0.5] -8.06068 -8.0714 -8.06723 -8.0691
LR_13[cache_size=0.2,treshold=0.6] -5.64059 -5.65901 -5.65255 -5.65466
LR_13[cache_size=0.2,treshold=0.7] -2.73336 -2.74608 -2.73908 -2.7398
LR_13[cache_size=0.2,treshold=0.8] 0 0 0 0
LR_13[cache_size=0.2,treshold=0.9] 0 0 0 0
LR_14[cache_size=0.2,treshold=0.3] -12.292 -12.2985 -12.2954 -12.2955
LR_14[cache_size=0.2,treshold=0.5] -8.05938 -8.07009 -8.06592 -8.06779
LR_14[cache_size=0.2,treshold=0.6] -5.64124 -5.65966 -5.65346 -5.65544
LR_14[cache_size=0.2,treshold=0.7] -2.73663 -2.75001 -2.74236 -2.74307
LR_14[cache_size=0.2,treshold=0.8] 0 0 0 0
LR_14[cache_size=0.2,treshold=0.9] 0 0 0 0
LR_15[cache_size=0.2,treshold=0.3] -12.3011 -12.3083 -12.3044 -12.304
LR_15[cache_size=0.2,treshold=0.5] -8.06265 -8.07336 -8.06933 -8.07106
LR_15[cache_size=0.2,treshold=0.6] -5.63993 -5.65835 -5.65202 -5.654
LR_15[cache_size=0.2,treshold=0.7] -2.72922 -2.74215 -2.73463 -2.73522
LR_15[cache_size=0.2,treshold=0.8] 0 0 0 0
LR_15[cache_size=0.2,treshold=0.9] 0 0 0 0
LR_1[cache_size=0.2,treshold=0.5] -4.34532 -4.3589 -4.3518 -4.35281
LR_1_log[cache_size=0.2,treshold=0.5] -4.49586 -4.51532 -4.50601 -4.50932
LR_1_mean[cache_size=0.2,treshold=0.5] -8.05283 -8.06224 -8.05794 -8.05928
LR_1_robust_scaler[cache_size=0.2,treshold=0.5] -4.34598 -4.3602 -4.35245 -4.35281
LR_1_std_scaler[cache_size=0.2,treshold=0.5] -4.34925 -4.36282 -4.3552 -4.35608
LR_2[cache_size=0.2,treshold=0.5] -6.12635 -6.15169 -6.14085 -6.14131
LR_2_log[cache_size=0.2,treshold=0.5] -8.0345 -8.04915 -8.04131 -8.04029
LR_2_mean[cache_size=0.2,treshold=0.5] -8.06134 -8.07205 -8.0671 -8.06845
LR_3[cache_size=0.2,treshold=0.5] -6.127 -6.15235 -6.14164 -6.14262
LR_3_log[cache_size=0.2,treshold=0.5] -8.03058 -8.04588 -8.03725 -8.03613
LR_3_mean[cache_size=0.2,treshold=0.5] -8.05676 -8.06747 -8.06252 -8.06386
LR_4[cache_size=0.2,treshold=0.5] -4.18562 -4.20051 -4.19182 -4.19302
LR_4_log[cache_size=0.2,treshold=0.5] -4.49324 -4.50943 -4.50038 -4.49624
LR_4_mean[cache_size=0.2,treshold=0.5] -7.44372 -7.45725 -7.45207 -7.45427
LR_4_robust_scaler[cache_size=0.2,treshold=0.5] -4.18562 -4.1992 -4.19169 -4.19237
LR_4_std_scaler[cache_size=0.2,treshold=0.5] -4.18627 -4.20116 -4.19287 -4.19368
LR_5[cache_size=0.2,treshold=0.5] -4.18955 -4.20378 -4.19562 -4.1963
LR_5_imba[cache_size=0.2,treshold=0.5] -2.44592 -2.46081 -2.45239 -2.45302
LR_6[cache_size=0.2,treshold=0.5] -4.18562 -4.1992 -4.19169 -4.19237
LR_6_imba[cache_size=0.2,treshold=0.5] -2.44199 -2.45623 -2.44872 -2.44975
LR_7[cache_size=0.2,treshold=0.3] -8.57414 -8.57858 -8.57596 -8.57545
LR_7[cache_size=0.2,treshold=0.5] -5.22041 -5.23591 -5.22656 -5.22357
LR_7[cache_size=0.2,treshold=0.6] -3.42769 -3.44523 -3.43567 -3.43798
LR_7[cache_size=0.2,treshold=0.7] -1.56102 -1.57424 -1.56833 -1.56881
LR_7[cache_size=0.2,treshold=0.8] -0.0824656 -0.0942717 -0.0864022 -0.0844324
LR_7[cache_size=0.2,treshold=0.9] -0.00196289 -0.00261941 -0.00209466 -0.00196354
LR_8[cache_size=0.2,treshold=0.3] -8.59705 -8.60215 -8.599 -8.59867
LR_8[cache_size=0.2,treshold=0.5] -5.2263 -5.24376 -5.23349 -5.23031
LR_8[cache_size=0.2,treshold=0.6] -3.42246 -3.43868 -3.42978 -3.43143
LR_8[cache_size=0.2,treshold=0.7] -1.54465 -1.55526 -1.55053 -1.55114
LR_8[cache_size=0.2,treshold=0.8] -0.0687596 -0.0792144 -0.0729181 -0.0719966
LR_8[cache_size=0.2,treshold=0.9] 0 0 0 0
LR_9[cache_size=0.2,treshold=0.3] -8.60164 -8.60739 -8.6041 -8.60359
LR_9[cache_size=0.2,treshold=0.5] -5.22892 -5.24507 -5.23546 -5.23359
LR_9[cache_size=0.2,treshold=0.6] -3.4218 -3.43868 -3.42939 -3.43078
LR_9[cache_size=0.2,treshold=0.7] -1.54138 -1.55264 -1.54791 -1.54852
LR_9[cache_size=0.2,treshold=0.8] -0.0674499 -0.0779051 -0.0718708 -0.0713421
LR_9[cache_size=0.2,treshold=0.9] 0 0 0 0
LR_10[cache_size=0.4,treshold=0.3] -14.2156 -14.275 -14.2511 -14.2538
LR_10[cache_size=0.4,treshold=0.5] -8.93748 -8.98745 -8.96031 -8.95992
LR_10[cache_size=0.4,treshold=0.6] -6.01236 -6.06569 -6.04422 -6.04908
LR_10[cache_size=0.4,treshold=0.7] -2.61501 -2.64685 -2.63335 -2.64157
LR_10[cache_size=0.4,treshold=0.8] 0 0 0 0
LR_10[cache_size=0.4,treshold=0.9] 0 0 0 0
LR_11[cache_size=0.4,treshold=0.3] -14.2123 -14.2728 -14.2484 -14.2516
LR_11[cache_size=0.4,treshold=0.5] -8.93726 -8.98711 -8.96007 -8.95981
LR_11[cache_size=0.4,treshold=0.6] -6.01314 -6.06703 -6.0452 -6.05019
LR_11[cache_size=0.4,treshold=0.7] -2.61635 -2.64852 -2.63487 -2.64336
LR_11[cache_size=0.4,treshold=0.8] 0 0 0 0
LR_11[cache_size=0.4,treshold=0.9] 0 0 0 0
LR_12[cache_size=0.4,treshold=0.3] -14.2179 -14.2784 -14.254 -14.2571
LR_12[cache_size=0.4,treshold=0.5] -8.93815 -8.98867 -8.96125 -8.96093
LR_12[cache_size=0.4,treshold=0.6] -6.01203 -6.06558 -6.0441 -6.0493
LR_12[cache_size=0.4,treshold=0.7] -2.61367 -2.64473 -2.63128 -2.63934
LR_12[cache_size=0.4,treshold=0.8] 0 -0.000111479 -2.22958e-05 0
LR_12[cache_size=0.4,treshold=0.9] 0 0 0 0
LR_13[cache_size=0.4,treshold=0.3] -24.7111 -24.7763 -24.7426 -24.7333
LR_13[cache_size=0.4,treshold=0.5] -18.4332 -18.4889 -18.4702 -18.4734
LR_13[cache_size=0.4,treshold=0.6] 0 0 0 0
LR_13[cache_size=0.4,treshold=0.7] 0 0 0 0
LR_13[cache_size=0.4,treshold=0.8] 0 0 0 0
LR_13[cache_size=0.4,treshold=0.9] 0 0 0 0
LR_14[cache_size=0.4,treshold=0.3] -24.7111 -24.7763 -24.7426 -24.7333
LR_14[cache_size=0.4,treshold=0.5] -18.4332 -18.4889 -18.4702 -18.4734
LR_14[cache_size=0.4,treshold=0.6] 0 0 0 0
LR_14[cache_size=0.4,treshold=0.7] 0 0 0 0
LR_14[cache_size=0.4,treshold=0.8] 0 0 0 0
LR_14[cache_size=0.4,treshold=0.9] 0 0 0 0
LR_15[cache_size=0.4,treshold=0.3] -24.7111 -24.7763 -24.7426 -24.7333
LR_15[cache_size=0.4,treshold=0.5] -18.4332 -18.4889 -18.4702 -18.4734
LR_15[cache_size=0.4,treshold=0.6] 0 0 0 0
LR_15[cache_size=0.4,treshold=0.7] 0 0 0 0
LR_15[cache_size=0.4,treshold=0.8] 0 0 0 0
LR_15[cache_size=0.4,treshold=0.9] 0 0 0 0
LR_1[cache_size=0.4,treshold=0.5] -4.27173 -4.33018 -4.29553 -4.28948
LR_1_log[cache_size=0.4,treshold=0.5] -4.87156 -4.92392 -4.89566 -4.89544
LR_1_mean[cache_size=0.4,treshold=0.5] -8.93002 -8.97831 -8.94945 -8.94771
LR_1_robust_scaler[cache_size=0.4,treshold=0.5] -4.27263 -4.32985 -4.29562 -4.28925
LR_1_std_scaler[cache_size=0.4,treshold=0.5] -4.28934 -4.34769 -4.31314 -4.30698
LR_2[cache_size=0.4,treshold=0.5] -8.08884 -8.14444 -8.11281 -8.11011
LR_2_log[cache_size=0.4,treshold=0.5] -8.47372 -8.52637 -8.49845 -8.49391
LR_2_mean[cache_size=0.4,treshold=0.5] -8.92723 -8.98198 -8.9513 -8.95072
LR_3[cache_size=0.4,treshold=0.5] -6.61497 -6.67559 -6.64662 -6.65118
LR_3_log[cache_size=0.4,treshold=0.5] -8.47405 -8.52804 -8.4993 -8.49514
LR_3_mean[cache_size=0.4,treshold=0.5] -8.92534 -8.97998 -8.94963 -8.94938
LR_4[cache_size=0.4,treshold=0.5] -4.48902 -4.54879 -4.51535 -4.50886
LR_4_log[cache_size=0.4,treshold=0.5] -4.8496 -4.90385 -4.87695 -4.87794
LR_4_mean[cache_size=0.4,treshold=0.5] -8.12873 -8.18992 -8.16157 -8.16416
LR_4_robust_scaler[cache_size=0.4,treshold=0.5] -4.20945 -4.26642 -4.23389 -4.22739
LR_4_std_scaler[cache_size=0.4,treshold=0.5] -4.22092 -4.27712 -4.24497 -4.23898
LR_5[cache_size=0.4,treshold=0.5] -4.22293 -4.27991 -4.24724 -4.24121
LR_5_imba[cache_size=0.4,treshold=0.5] -1.87457 -1.92023 -1.90385 -1.90786
LR_6[cache_size=0.4,treshold=0.5] -4.20755 -4.2653 -4.23215 -4.22583
LR_6_imba[cache_size=0.4,treshold=0.5] -1.86677 -1.91856 -1.89866 -1.90084
LR_7[cache_size=0.4,treshold=0.3] -9.46621 -9.52188 -9.49432 -9.49426
LR_7[cache_size=0.4,treshold=0.5] -5.35293 -5.42246 -5.39159 -5.38726
LR_7[cache_size=0.4,treshold=0.6] -3.23244 -3.29443 -3.26094 -3.25586
LR_7[cache_size=0.4,treshold=0.7] -1.19619 -1.23474 -1.22055 -1.22163
LR_7[cache_size=0.4,treshold=0.8] -0.102515 -0.129588 -0.117868 -0.122509
LR_7[cache_size=0.4,treshold=0.9] 0 0 0 0
LR_8[cache_size=0.4,treshold=0.3] -9.46543 -9.52076 -9.4928 -9.49181
LR_8[cache_size=0.4,treshold=0.5] -5.35315 -5.42279 -5.39183 -5.38793
LR_8[cache_size=0.4,treshold=0.6] -3.23378 -3.29477 -3.26206 -3.25686
LR_8[cache_size=0.4,treshold=0.7] -1.19797 -1.2363 -1.22197 -1.2223
LR_8[cache_size=0.4,treshold=0.8] -0.103295 -0.130034 -0.118113 -0.122286
LR_8[cache_size=0.4,treshold=0.9] 0 0 0 0
LR_9[cache_size=0.4,treshold=0.3] -9.46666 -9.52165 -9.49456 -9.49459
LR_9[cache_size=0.4,treshold=0.5] -5.35326 -5.42312 -5.39172 -5.38693
LR_9[cache_size=0.4,treshold=0.6] -3.23233 -3.29443 -3.2609 -3.25575
LR_9[cache_size=0.4,treshold=0.7] -1.19608 -1.23507 -1.22055 -1.2213
LR_9[cache_size=0.4,treshold=0.8] -0.103183 -0.129477 -0.118047 -0.122509
LR_9[cache_size=0.4,treshold=0.9] 0 0 0 0
Offline Clock 1st iteration 0 0 0 0
Offline Clock 2nd iteration 7.80369 0.100977 2.73498 1.84941
Zipf Optimal Distribution 3.35401 0.0684004 1.23463 0.902408

Model Summaries Plot

Miss Ratio Reduced (%)

Promotion Reduced (%)

Promotion vs Miss Ratio

Cache Size All

Cache Size 0.001

Cache Size 0.01

Cache Size 0.1

Cache Size 0.2

Cache Size 0.4

Model Classification Report

LR_10_0.001

              precision    recall  f1-score   support

           0       0.89      0.71      0.79  30469299
           1       0.66      0.87      0.75  19888532

    accuracy                           0.77  50357831
   macro avg       0.78      0.79      0.77  50357831
weighted avg       0.80      0.77      0.78  50357831

Accuracy: 0.7738074342399696
              precision    recall  f1-score   support

           0       0.89      0.71      0.79  30469299
           1       0.66      0.87      0.75  19888532

    accuracy                           0.77  50357831
   macro avg       0.78      0.79      0.77  50357831
weighted avg       0.80      0.77      0.78  50357831

Accuracy: 0.7738074342399696

LR_10_0.01

              precision    recall  f1-score   support

           0       0.89      0.69      0.78  30231476
           1       0.65      0.87      0.74  19809713

    accuracy                           0.76  50041189
   macro avg       0.77      0.78      0.76  50041189
weighted avg       0.79      0.76      0.76  50041189

Accuracy: 0.7611738202303706
              precision    recall  f1-score   support

           0       0.89      0.69      0.78  30231476
           1       0.65      0.87      0.74  19809713

    accuracy                           0.76  50041189
   macro avg       0.77      0.78      0.76  50041189
weighted avg       0.79      0.76      0.76  50041189

Accuracy: 0.7611738202303706

LR_10_0.1

              precision    recall  f1-score   support

           0       0.87      0.59      0.70  29263977
           1       0.58      0.86      0.69  18981990

    accuracy                           0.70  48245967
   macro avg       0.72      0.73      0.70  48245967
weighted avg       0.76      0.70      0.70  48245967

Accuracy: 0.698262447511934
              precision    recall  f1-score   support

           0       0.87      0.59      0.70  29263977
           1       0.58      0.86      0.69  18981990

    accuracy                           0.70  48245967
   macro avg       0.72      0.73      0.70  48245967
weighted avg       0.76      0.70      0.70  48245967

Accuracy: 0.698262447511934

LR_10_0.2

              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.82      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.73      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7202710005355438
              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.82      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.73      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7202710005355438

LR_10_0.4

              precision    recall  f1-score   support

           0       0.89      0.69      0.77  26121772
           1       0.60      0.84      0.70  14505950

    accuracy                           0.74  40627722
   macro avg       0.74      0.77      0.74  40627722
weighted avg       0.78      0.74      0.75  40627722

Accuracy: 0.7429382331601068
              precision    recall  f1-score   support

           0       0.89      0.69      0.77  26121772
           1       0.60      0.84      0.70  14505950

    accuracy                           0.74  40627722
   macro avg       0.74      0.77      0.74  40627722
weighted avg       0.78      0.74      0.75  40627722

Accuracy: 0.7429382331601068

LR_10_All

              precision    recall  f1-score   support

           0       0.73      0.31      0.44 144486817
           1       0.43      0.81      0.56  90948310

    accuracy                           0.51 235435127
   macro avg       0.58      0.56      0.50 235435127
weighted avg       0.61      0.51      0.48 235435127

Accuracy: 0.5060794899989584
              precision    recall  f1-score   support

           0       0.73      0.31      0.44 144486817
           1       0.43      0.81      0.56  90948310

    accuracy                           0.51 235435127
   macro avg       0.58      0.56      0.50 235435127
weighted avg       0.61      0.51      0.48 235435127

Accuracy: 0.5060794899989584

LR_11_0.001

              precision    recall  f1-score   support

           0       0.89      0.71      0.79  30469299
           1       0.66      0.86      0.75  19888532

    accuracy                           0.77  50357831
   macro avg       0.77      0.79      0.77  50357831
weighted avg       0.80      0.77      0.77  50357831

Accuracy: 0.7689939822864889
              precision    recall  f1-score   support

           0       0.89      0.71      0.79  30469299
           1       0.66      0.86      0.75  19888532

    accuracy                           0.77  50357831
   macro avg       0.77      0.79      0.77  50357831
weighted avg       0.80      0.77      0.77  50357831

Accuracy: 0.7689939822864889

LR_11_0.01

              precision    recall  f1-score   support

           0       0.89      0.69      0.78  30231476
           1       0.65      0.87      0.74  19809713

    accuracy                           0.76  50041189
   macro avg       0.77      0.78      0.76  50041189
weighted avg       0.79      0.76      0.76  50041189

Accuracy: 0.7614748322626786
              precision    recall  f1-score   support

           0       0.89      0.69      0.78  30231476
           1       0.65      0.87      0.74  19809713

    accuracy                           0.76  50041189
   macro avg       0.77      0.78      0.76  50041189
weighted avg       0.79      0.76      0.76  50041189

Accuracy: 0.7614748322626786

LR_11_0.1

              precision    recall  f1-score   support

           0       0.87      0.59      0.70  29263977
           1       0.58      0.87      0.69  18981990

    accuracy                           0.70  48245967
   macro avg       0.73      0.73      0.70  48245967
weighted avg       0.76      0.70      0.70  48245967

Accuracy: 0.6982573486401464
              precision    recall  f1-score   support

           0       0.87      0.59      0.70  29263977
           1       0.58      0.87      0.69  18981990

    accuracy                           0.70  48245967
   macro avg       0.73      0.73      0.70  48245967
weighted avg       0.76      0.70      0.70  48245967

Accuracy: 0.6982573486401464

LR_11_0.2

              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.82      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.73      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7202601258885529
              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.82      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.73      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7202601258885529

LR_11_0.4

              precision    recall  f1-score   support

           0       0.89      0.69      0.77  26121772
           1       0.60      0.84      0.70  14505950

    accuracy                           0.74  40627722
   macro avg       0.74      0.77      0.74  40627722
weighted avg       0.78      0.74      0.75  40627722

Accuracy: 0.742940891443532
              precision    recall  f1-score   support

           0       0.89      0.69      0.77  26121772
           1       0.60      0.84      0.70  14505950

    accuracy                           0.74  40627722
   macro avg       0.74      0.77      0.74  40627722
weighted avg       0.78      0.74      0.75  40627722

Accuracy: 0.742940891443532

LR_11_All

              precision    recall  f1-score   support

           0       0.73      0.31      0.44 144486817
           1       0.43      0.81      0.56  90948310

    accuracy                           0.51 235435127
   macro avg       0.58      0.56      0.50 235435127
weighted avg       0.61      0.51      0.48 235435127

Accuracy: 0.5060570082220569
              precision    recall  f1-score   support

           0       0.73      0.31      0.44 144486817
           1       0.43      0.81      0.56  90948310

    accuracy                           0.51 235435127
   macro avg       0.58      0.56      0.50 235435127
weighted avg       0.61      0.51      0.48 235435127

Accuracy: 0.5060570082220569

LR_12_0.001

              precision    recall  f1-score   support

           0       0.89      0.71      0.79  30469299
           1       0.66      0.87      0.75  19888532

    accuracy                           0.77  50357831
   macro avg       0.78      0.79      0.77  50357831
weighted avg       0.80      0.77      0.78  50357831

Accuracy: 0.7734336691347965
              precision    recall  f1-score   support

           0       0.89      0.71      0.79  30469299
           1       0.66      0.87      0.75  19888532

    accuracy                           0.77  50357831
   macro avg       0.78      0.79      0.77  50357831
weighted avg       0.80      0.77      0.78  50357831

Accuracy: 0.7734336691347965

LR_12_0.01

              precision    recall  f1-score   support

           0       0.89      0.69      0.78  30231476
           1       0.65      0.87      0.74  19809713

    accuracy                           0.76  50041189
   macro avg       0.77      0.78      0.76  50041189
weighted avg       0.79      0.76      0.76  50041189

Accuracy: 0.7608722086919237
              precision    recall  f1-score   support

           0       0.89      0.69      0.78  30231476
           1       0.65      0.87      0.74  19809713

    accuracy                           0.76  50041189
   macro avg       0.77      0.78      0.76  50041189
weighted avg       0.79      0.76      0.76  50041189

Accuracy: 0.7608722086919237

LR_12_0.1

              precision    recall  f1-score   support

           0       0.88      0.59      0.70  29263977
           1       0.58      0.87      0.69  18981990

    accuracy                           0.70  48245967
   macro avg       0.73      0.73      0.70  48245967
weighted avg       0.76      0.70      0.70  48245967

Accuracy: 0.6982929785612961
              precision    recall  f1-score   support

           0       0.88      0.59      0.70  29263977
           1       0.58      0.87      0.69  18981990

    accuracy                           0.70  48245967
   macro avg       0.73      0.73      0.70  48245967
weighted avg       0.76      0.70      0.70  48245967

Accuracy: 0.6982929785612961

LR_12_0.2

              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.82      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.73      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7202365352698812
              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.82      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.73      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7202365352698812

LR_12_0.4

              precision    recall  f1-score   support

           0       0.89      0.69      0.77  26121772
           1       0.60      0.84      0.70  14505950

    accuracy                           0.74  40627722
   macro avg       0.74      0.77      0.74  40627722
weighted avg       0.78      0.74      0.75  40627722

Accuracy: 0.7429330642756687
              precision    recall  f1-score   support

           0       0.89      0.69      0.77  26121772
           1       0.60      0.84      0.70  14505950

    accuracy                           0.74  40627722
   macro avg       0.74      0.77      0.74  40627722
weighted avg       0.78      0.74      0.75  40627722

Accuracy: 0.7429330642756687

LR_12_All

              precision    recall  f1-score   support

           0       0.73      0.31      0.44 144486817
           1       0.43      0.81      0.56  90948310

    accuracy                           0.51 235435127
   macro avg       0.58      0.56      0.50 235435127
weighted avg       0.61      0.51      0.48 235435127

Accuracy: 0.506046695402424
              precision    recall  f1-score   support

           0       0.73      0.31      0.44 144486817
           1       0.43      0.81      0.56  90948310

    accuracy                           0.51 235435127
   macro avg       0.58      0.56      0.50 235435127
weighted avg       0.61      0.51      0.48 235435127

Accuracy: 0.506046695402424

LR_13_0.001

              precision    recall  f1-score   support

           0       0.89      0.69      0.78  30469299
           1       0.65      0.87      0.74  19888532

    accuracy                           0.76  50357831
   macro avg       0.77      0.78      0.76  50357831
weighted avg       0.79      0.76      0.76  50357831

Accuracy: 0.7611932094533619
              precision    recall  f1-score   support

           0       0.89      0.69      0.78  30469299
           1       0.65      0.87      0.74  19888532

    accuracy                           0.76  50357831
   macro avg       0.77      0.78      0.76  50357831
weighted avg       0.79      0.76      0.76  50357831

Accuracy: 0.7611932094533619

LR_13_0.01

              precision    recall  f1-score   support

           0       0.87      0.66      0.75  30231476
           1       0.62      0.85      0.72  19809713

    accuracy                           0.74  50041189
   macro avg       0.75      0.76      0.74  50041189
weighted avg       0.77      0.74      0.74  50041189

Accuracy: 0.7379024507191466
              precision    recall  f1-score   support

           0       0.87      0.66      0.75  30231476
           1       0.62      0.85      0.72  19809713

    accuracy                           0.74  50041189
   macro avg       0.75      0.76      0.74  50041189
weighted avg       0.77      0.74      0.74  50041189

Accuracy: 0.7379024507191466

LR_13_0.1

              precision    recall  f1-score   support

           0       0.83      0.64      0.72  29263977
           1       0.59      0.80      0.68  18981990

    accuracy                           0.70  48245967
   macro avg       0.71      0.72      0.70  48245967
weighted avg       0.74      0.70      0.71  48245967

Accuracy: 0.7041242431724914
              precision    recall  f1-score   support

           0       0.83      0.64      0.72  29263977
           1       0.59      0.80      0.68  18981990

    accuracy                           0.70  48245967
   macro avg       0.71      0.72      0.70  48245967
weighted avg       0.74      0.70      0.71  48245967

Accuracy: 0.7041242431724914

LR_13_0.2

              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.82      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.73      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7202479514829574
              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.82      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.73      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7202479514829574

LR_13_0.4

              precision    recall  f1-score   support

           0       0.89      0.66      0.76  26121772
           1       0.58      0.86      0.70  14505950

    accuracy                           0.73  40627722
   macro avg       0.74      0.76      0.73  40627722
weighted avg       0.78      0.73      0.74  40627722

Accuracy: 0.7319589811114686
              precision    recall  f1-score   support

           0       0.89      0.66      0.76  26121772
           1       0.58      0.86      0.70  14505950

    accuracy                           0.73  40627722
   macro avg       0.74      0.76      0.73  40627722
weighted avg       0.78      0.73      0.74  40627722

Accuracy: 0.7319589811114686

LR_13_All

              precision    recall  f1-score   support

           0       0.69      0.31      0.43 144486817
           1       0.41      0.77      0.54  90948310

    accuracy                           0.49 235435127
   macro avg       0.55      0.54      0.49 235435127
weighted avg       0.58      0.49      0.47 235435127

Accuracy: 0.49110166555562457
              precision    recall  f1-score   support

           0       0.69      0.31      0.43 144486817
           1       0.41      0.77      0.54  90948310

    accuracy                           0.49 235435127
   macro avg       0.55      0.54      0.49 235435127
weighted avg       0.58      0.49      0.47 235435127

Accuracy: 0.49110166555562457

LR_14_0.001

              precision    recall  f1-score   support

           0       0.88      0.69      0.77  30469299
           1       0.64      0.86      0.74  19888532

    accuracy                           0.76  50357831
   macro avg       0.76      0.77      0.75  50357831
weighted avg       0.79      0.76      0.76  50357831

Accuracy: 0.7556111779317898
              precision    recall  f1-score   support

           0       0.88      0.69      0.77  30469299
           1       0.64      0.86      0.74  19888532

    accuracy                           0.76  50357831
   macro avg       0.76      0.77      0.75  50357831
weighted avg       0.79      0.76      0.76  50357831

Accuracy: 0.7556111779317898

LR_14_0.01

              precision    recall  f1-score   support

           0       0.87      0.66      0.75  30231476
           1       0.62      0.85      0.72  19809713

    accuracy                           0.74  50041189
   macro avg       0.75      0.76      0.74  50041189
weighted avg       0.77      0.74      0.74  50041189

Accuracy: 0.7379092651055913
              precision    recall  f1-score   support

           0       0.87      0.66      0.75  30231476
           1       0.62      0.85      0.72  19809713

    accuracy                           0.74  50041189
   macro avg       0.75      0.76      0.74  50041189
weighted avg       0.77      0.74      0.74  50041189

Accuracy: 0.7379092651055913

LR_14_0.1

              precision    recall  f1-score   support

           0       0.83      0.64      0.72  29263977
           1       0.59      0.80      0.68  18981990

    accuracy                           0.70  48245967
   macro avg       0.71      0.72      0.70  48245967
weighted avg       0.74      0.70      0.71  48245967

Accuracy: 0.7040489622687012
              precision    recall  f1-score   support

           0       0.83      0.64      0.72  29263977
           1       0.59      0.80      0.68  18981990

    accuracy                           0.70  48245967
   macro avg       0.71      0.72      0.70  48245967
weighted avg       0.74      0.70      0.71  48245967

Accuracy: 0.7040489622687012

LR_14_0.2

              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.82      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.73      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7202581329253593
              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.82      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.73      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7202581329253593

LR_14_0.4

              precision    recall  f1-score   support

           0       0.90      0.63      0.74  26121772
           1       0.57      0.87      0.69  14505950

    accuracy                           0.72  40627722
   macro avg       0.73      0.75      0.71  40627722
weighted avg       0.78      0.72      0.72  40627722

Accuracy: 0.7152666595483744
              precision    recall  f1-score   support

           0       0.90      0.63      0.74  26121772
           1       0.57      0.87      0.69  14505950

    accuracy                           0.72  40627722
   macro avg       0.73      0.75      0.71  40627722
weighted avg       0.78      0.72      0.72  40627722

Accuracy: 0.7152666595483744

LR_14_All

              precision    recall  f1-score   support

           0       0.69      0.32      0.44 144486817
           1       0.42      0.77      0.54  90948310

    accuracy                           0.49 235435127
   macro avg       0.55      0.55      0.49 235435127
weighted avg       0.58      0.49      0.48 235435127

Accuracy: 0.4946263604920773
              precision    recall  f1-score   support

           0       0.69      0.32      0.44 144486817
           1       0.42      0.77      0.54  90948310

    accuracy                           0.49 235435127
   macro avg       0.55      0.55      0.49 235435127
weighted avg       0.58      0.49      0.48 235435127

Accuracy: 0.4946263604920773

LR_15_0.001

              precision    recall  f1-score   support

           0       0.89      0.69      0.78  30469299
           1       0.65      0.86      0.74  19888532

    accuracy                           0.76  50357831
   macro avg       0.77      0.78      0.76  50357831
weighted avg       0.79      0.76      0.76  50357831

Accuracy: 0.7611176899179792
              precision    recall  f1-score   support

           0       0.89      0.69      0.78  30469299
           1       0.65      0.86      0.74  19888532

    accuracy                           0.76  50357831
   macro avg       0.77      0.78      0.76  50357831
weighted avg       0.79      0.76      0.76  50357831

Accuracy: 0.7611176899179792

LR_15_0.01

              precision    recall  f1-score   support

           0       0.87      0.66      0.75  30231476
           1       0.62      0.85      0.72  19809713

    accuracy                           0.74  50041189
   macro avg       0.75      0.76      0.74  50041189
weighted avg       0.77      0.74      0.74  50041189

Accuracy: 0.7379364227336804
              precision    recall  f1-score   support

           0       0.87      0.66      0.75  30231476
           1       0.62      0.85      0.72  19809713

    accuracy                           0.74  50041189
   macro avg       0.75      0.76      0.74  50041189
weighted avg       0.77      0.74      0.74  50041189

Accuracy: 0.7379364227336804

LR_15_0.1

              precision    recall  f1-score   support

           0       0.83      0.64      0.72  29263977
           1       0.59      0.80      0.68  18981990

    accuracy                           0.70  48245967
   macro avg       0.71      0.72      0.70  48245967
weighted avg       0.74      0.70      0.71  48245967

Accuracy: 0.7039353569180197
              precision    recall  f1-score   support

           0       0.83      0.64      0.72  29263977
           1       0.59      0.80      0.68  18981990

    accuracy                           0.70  48245967
   macro avg       0.71      0.72      0.70  48245967
weighted avg       0.74      0.70      0.71  48245967

Accuracy: 0.7039353569180197

LR_15_0.2

              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.82      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.73      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7202286933929675
              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.82      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.73      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7202286933929675

LR_15_0.4

              precision    recall  f1-score   support

           0       0.90      0.63      0.74  26121772
           1       0.57      0.87      0.69  14505950

    accuracy                           0.72  40627722
   macro avg       0.73      0.75      0.71  40627722
weighted avg       0.78      0.72      0.72  40627722

Accuracy: 0.7152346124648583
              precision    recall  f1-score   support

           0       0.90      0.63      0.74  26121772
           1       0.57      0.87      0.69  14505950

    accuracy                           0.72  40627722
   macro avg       0.73      0.75      0.71  40627722
weighted avg       0.78      0.72      0.72  40627722

Accuracy: 0.7152346124648583

LR_15_All

              precision    recall  f1-score   support

           0       0.69      0.32      0.44 144486817
           1       0.42      0.77      0.54  90948310

    accuracy                           0.49 235435127
   macro avg       0.55      0.55      0.49 235435127
weighted avg       0.58      0.49      0.48 235435127

Accuracy: 0.49461241609880924
              precision    recall  f1-score   support

           0       0.69      0.32      0.44 144486817
           1       0.42      0.77      0.54  90948310

    accuracy                           0.49 235435127
   macro avg       0.55      0.55      0.49 235435127
weighted avg       0.58      0.49      0.48 235435127

Accuracy: 0.49461241609880924

LR_1_0.001

              precision    recall  f1-score   support

           0       0.86      0.66      0.75  30469299
           1       0.62      0.83      0.71  19888532

    accuracy                           0.73  50357831
   macro avg       0.74      0.75      0.73  50357831
weighted avg       0.76      0.73      0.73  50357831

Accuracy: 0.7289517890474672
              precision    recall  f1-score   support

           0       0.86      0.66      0.75  30469299
           1       0.62      0.83      0.71  19888532

    accuracy                           0.73  50357831
   macro avg       0.74      0.75      0.73  50357831
weighted avg       0.76      0.73      0.73  50357831

Accuracy: 0.7289517890474672

LR_1_0.01

              precision    recall  f1-score   support

           0       0.86      0.68      0.76  30231476
           1       0.63      0.83      0.72  19809713

    accuracy                           0.74  50041189
   macro avg       0.74      0.76      0.74  50041189
weighted avg       0.77      0.74      0.74  50041189

Accuracy: 0.7399302402666731
              precision    recall  f1-score   support

           0       0.86      0.68      0.76  30231476
           1       0.63      0.83      0.72  19809713

    accuracy                           0.74  50041189
   macro avg       0.74      0.76      0.74  50041189
weighted avg       0.77      0.74      0.74  50041189

Accuracy: 0.7399302402666731

LR_1_0.1

              precision    recall  f1-score   support

           0       0.86      0.73      0.79  29263977
           1       0.67      0.82      0.74  18981990

    accuracy                           0.77  48245967
   macro avg       0.76      0.78      0.76  48245967
weighted avg       0.79      0.77      0.77  48245967

Accuracy: 0.7676230637060296
              precision    recall  f1-score   support

           0       0.86      0.73      0.79  29263977
           1       0.67      0.82      0.74  18981990

    accuracy                           0.77  48245967
   macro avg       0.76      0.78      0.76  48245967
weighted avg       0.79      0.77      0.77  48245967

Accuracy: 0.7676230637060296

LR_1_0.2

              precision    recall  f1-score   support

           0       0.87      0.75      0.81  28400293
           1       0.67      0.81      0.74  17762125

    accuracy                           0.78  46162418
   macro avg       0.77      0.78      0.77  46162418
weighted avg       0.79      0.78      0.78  46162418

Accuracy: 0.7765185523860557
              precision    recall  f1-score   support

           0       0.87      0.75      0.81  28400293
           1       0.67      0.81      0.74  17762125

    accuracy                           0.78  46162418
   macro avg       0.77      0.78      0.77  46162418
weighted avg       0.79      0.78      0.78  46162418

Accuracy: 0.7765185523860557

LR_1_0.4

              precision    recall  f1-score   support

           0       0.87      0.77      0.82  26121772
           1       0.66      0.80      0.72  14505950

    accuracy                           0.78  40627722
   macro avg       0.76      0.78      0.77  40627722
weighted avg       0.79      0.78      0.78  40627722

Accuracy: 0.7780763587975718
              precision    recall  f1-score   support

           0       0.87      0.77      0.82  26121772
           1       0.66      0.80      0.72  14505950

    accuracy                           0.78  40627722
   macro avg       0.76      0.78      0.77  40627722
weighted avg       0.79      0.78      0.78  40627722

Accuracy: 0.7780763587975718

LR_1_All

              precision    recall  f1-score   support

           0       0.67      0.60      0.63 144486817
           1       0.46      0.54      0.50  90948310

    accuracy                           0.58 235435127
   macro avg       0.57      0.57      0.57 235435127
weighted avg       0.59      0.58      0.58 235435127

Accuracy: 0.5767005872492426
              precision    recall  f1-score   support

           0       0.67      0.60      0.63 144486817
           1       0.46      0.54      0.50  90948310

    accuracy                           0.58 235435127
   macro avg       0.57      0.57      0.57 235435127
weighted avg       0.59      0.58      0.58 235435127

Accuracy: 0.5767005872492426

LR_1_log_0.001

              precision    recall  f1-score   support

           0       0.83      0.73      0.78  30469299
           1       0.65      0.77      0.71  19888532

    accuracy                           0.75  50357831
   macro avg       0.74      0.75      0.74  50357831
weighted avg       0.76      0.75      0.75  50357831

Accuracy: 0.7465433131939301
              precision    recall  f1-score   support

           0       0.83      0.73      0.78  30469299
           1       0.65      0.77      0.71  19888532

    accuracy                           0.75  50357831
   macro avg       0.74      0.75      0.74  50357831
weighted avg       0.76      0.75      0.75  50357831

Accuracy: 0.7465433131939301

LR_1_log_0.01

              precision    recall  f1-score   support

           0       0.84      0.73      0.78  30231476
           1       0.66      0.78      0.71  19809713

    accuracy                           0.75  50041189
   macro avg       0.75      0.76      0.75  50041189
weighted avg       0.77      0.75      0.75  50041189

Accuracy: 0.751373373642261
              precision    recall  f1-score   support

           0       0.84      0.73      0.78  30231476
           1       0.66      0.78      0.71  19809713

    accuracy                           0.75  50041189
   macro avg       0.75      0.76      0.75  50041189
weighted avg       0.77      0.75      0.75  50041189

Accuracy: 0.751373373642261

LR_1_log_0.1

              precision    recall  f1-score   support

           0       0.86      0.73      0.79  29263977
           1       0.66      0.81      0.73  18981990

    accuracy                           0.76  48245967
   macro avg       0.76      0.77      0.76  48245967
weighted avg       0.78      0.76      0.76  48245967

Accuracy: 0.761670193075413
              precision    recall  f1-score   support

           0       0.86      0.73      0.79  29263977
           1       0.66      0.81      0.73  18981990

    accuracy                           0.76  48245967
   macro avg       0.76      0.77      0.76  48245967
weighted avg       0.78      0.76      0.76  48245967

Accuracy: 0.761670193075413

LR_1_log_0.2

              precision    recall  f1-score   support

           0       0.87      0.72      0.79  28400293
           1       0.65      0.83      0.73  17762125

    accuracy                           0.76  46162418
   macro avg       0.76      0.77      0.76  46162418
weighted avg       0.79      0.76      0.77  46162418

Accuracy: 0.7624651507639829
              precision    recall  f1-score   support

           0       0.87      0.72      0.79  28400293
           1       0.65      0.83      0.73  17762125

    accuracy                           0.76  46162418
   macro avg       0.76      0.77      0.76  46162418
weighted avg       0.79      0.76      0.77  46162418

Accuracy: 0.7624651507639829

LR_1_log_0.4

              precision    recall  f1-score   support

           0       0.88      0.71      0.79  26121772
           1       0.61      0.83      0.71  14505950

    accuracy                           0.75  40627722
   macro avg       0.75      0.77      0.75  40627722
weighted avg       0.79      0.75      0.76  40627722

Accuracy: 0.7535204164289595
              precision    recall  f1-score   support

           0       0.88      0.71      0.79  26121772
           1       0.61      0.83      0.71  14505950

    accuracy                           0.75  40627722
   macro avg       0.75      0.77      0.75  40627722
weighted avg       0.79      0.75      0.76  40627722

Accuracy: 0.7535204164289595

LR_1_log_All

              precision    recall  f1-score   support

           0       0.73      0.56      0.64 144486817
           1       0.49      0.67      0.57  90948310

    accuracy                           0.61 235435127
   macro avg       0.61      0.62      0.60 235435127
weighted avg       0.64      0.61      0.61 235435127

Accuracy: 0.6053789670901573
              precision    recall  f1-score   support

           0       0.73      0.56      0.64 144486817
           1       0.49      0.67      0.57  90948310

    accuracy                           0.61 235435127
   macro avg       0.61      0.62      0.60 235435127
weighted avg       0.64      0.61      0.61 235435127

Accuracy: 0.6053789670901573

LR_1_mean_0.001

              precision    recall  f1-score   support

           0       0.86      0.71      0.78  30469299
           1       0.65      0.82      0.73  19888532

    accuracy                           0.75  50357831
   macro avg       0.75      0.77      0.75  50357831
weighted avg       0.78      0.75      0.76  50357831

Accuracy: 0.7549619243926531
              precision    recall  f1-score   support

           0       0.86      0.71      0.78  30469299
           1       0.65      0.82      0.73  19888532

    accuracy                           0.75  50357831
   macro avg       0.75      0.77      0.75  50357831
weighted avg       0.78      0.75      0.76  50357831

Accuracy: 0.7549619243926531

LR_1_mean_0.01

              precision    recall  f1-score   support

           0       0.81      0.66      0.73  30231476
           1       0.60      0.76      0.67  19809713

    accuracy                           0.70  50041189
   macro avg       0.70      0.71      0.70  50041189
weighted avg       0.73      0.70      0.71  50041189

Accuracy: 0.7019766456788227
              precision    recall  f1-score   support

           0       0.81      0.66      0.73  30231476
           1       0.60      0.76      0.67  19809713

    accuracy                           0.70  50041189
   macro avg       0.70      0.71      0.70  50041189
weighted avg       0.73      0.70      0.71  50041189

Accuracy: 0.7019766456788227

LR_1_mean_0.1

              precision    recall  f1-score   support

           0       0.83      0.64      0.72  29263977
           1       0.59      0.80      0.68  18981990

    accuracy                           0.70  48245967
   macro avg       0.71      0.72      0.70  48245967
weighted avg       0.74      0.70      0.71  48245967

Accuracy: 0.7043372143416672
              precision    recall  f1-score   support

           0       0.83      0.64      0.72  29263977
           1       0.59      0.80      0.68  18981990

    accuracy                           0.70  48245967
   macro avg       0.71      0.72      0.70  48245967
weighted avg       0.74      0.70      0.71  48245967

Accuracy: 0.7043372143416672

LR_1_mean_0.2

              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.82      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.73      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7203463865346048
              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.82      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.73      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7203463865346048

LR_1_mean_0.4

              precision    recall  f1-score   support

           0       0.89      0.69      0.77  26121772
           1       0.60      0.84      0.70  14505950

    accuracy                           0.74  40627722
   macro avg       0.74      0.76      0.74  40627722
weighted avg       0.78      0.74      0.75  40627722

Accuracy: 0.7424505612202427
              precision    recall  f1-score   support

           0       0.89      0.69      0.77  26121772
           1       0.60      0.84      0.70  14505950

    accuracy                           0.74  40627722
   macro avg       0.74      0.76      0.74  40627722
weighted avg       0.78      0.74      0.75  40627722

Accuracy: 0.7424505612202427

LR_1_mean_All

              precision    recall  f1-score   support

           0       0.69      0.39      0.50 144486817
           1       0.42      0.72      0.53  90948310

    accuracy                           0.52 235435127
   macro avg       0.55      0.55      0.51 235435127
weighted avg       0.58      0.52      0.51 235435127

Accuracy: 0.5151163105750146
              precision    recall  f1-score   support

           0       0.69      0.39      0.50 144486817
           1       0.42      0.72      0.53  90948310

    accuracy                           0.52 235435127
   macro avg       0.55      0.55      0.51 235435127
weighted avg       0.58      0.52      0.51 235435127

Accuracy: 0.5151163105750146

LR_1_robust_scaler_0.001

              precision    recall  f1-score   support

           0       0.86      0.66      0.75  30469299
           1       0.62      0.83      0.71  19888532

    accuracy                           0.73  50357831
   macro avg       0.74      0.75      0.73  50357831
weighted avg       0.76      0.73      0.73  50357831

Accuracy: 0.7289659477192336
              precision    recall  f1-score   support

           0       0.86      0.66      0.75  30469299
           1       0.62      0.83      0.71  19888532

    accuracy                           0.73  50357831
   macro avg       0.74      0.75      0.73  50357831
weighted avg       0.76      0.73      0.73  50357831

Accuracy: 0.7289659477192336

LR_1_robust_scaler_0.01

              precision    recall  f1-score   support

           0       0.86      0.68      0.76  30231476
           1       0.63      0.83      0.72  19809713

    accuracy                           0.74  50041189
   macro avg       0.74      0.76      0.74  50041189
weighted avg       0.77      0.74      0.74  50041189

Accuracy: 0.7398954289435449
              precision    recall  f1-score   support

           0       0.86      0.68      0.76  30231476
           1       0.63      0.83      0.72  19809713

    accuracy                           0.74  50041189
   macro avg       0.74      0.76      0.74  50041189
weighted avg       0.77      0.74      0.74  50041189

Accuracy: 0.7398954289435449

LR_1_robust_scaler_0.1

              precision    recall  f1-score   support

           0       0.86      0.73      0.79  29263977
           1       0.67      0.82      0.74  18981990

    accuracy                           0.77  48245967
   macro avg       0.76      0.78      0.76  48245967
weighted avg       0.79      0.77      0.77  48245967

Accuracy: 0.7676340283530849
              precision    recall  f1-score   support

           0       0.86      0.73      0.79  29263977
           1       0.67      0.82      0.74  18981990

    accuracy                           0.77  48245967
   macro avg       0.76      0.78      0.76  48245967
weighted avg       0.79      0.77      0.77  48245967

Accuracy: 0.7676340283530849

LR_1_robust_scaler_0.2

              precision    recall  f1-score   support

           0       0.87      0.75      0.81  28400293
           1       0.67      0.81      0.74  17762125

    accuracy                           0.78  46162418
   macro avg       0.77      0.78      0.77  46162418
weighted avg       0.79      0.78      0.78  46162418

Accuracy: 0.7765035618368171
              precision    recall  f1-score   support

           0       0.87      0.75      0.81  28400293
           1       0.67      0.81      0.74  17762125

    accuracy                           0.78  46162418
   macro avg       0.77      0.78      0.77  46162418
weighted avg       0.79      0.78      0.78  46162418

Accuracy: 0.7765035618368171

LR_1_robust_scaler_0.4

              precision    recall  f1-score   support

           0       0.87      0.77      0.82  26121772
           1       0.66      0.80      0.72  14505950

    accuracy                           0.78  40627722
   macro avg       0.76      0.78      0.77  40627722
weighted avg       0.79      0.78      0.78  40627722

Accuracy: 0.7780789678535263
              precision    recall  f1-score   support

           0       0.87      0.77      0.82  26121772
           1       0.66      0.80      0.72  14505950

    accuracy                           0.78  40627722
   macro avg       0.76      0.78      0.77  40627722
weighted avg       0.79      0.78      0.78  40627722

Accuracy: 0.7780789678535263

LR_1_robust_scaler_All

              precision    recall  f1-score   support

           0       0.69      0.51      0.59 144486817
           1       0.45      0.63      0.52  90948310

    accuracy                           0.56 235435127
   macro avg       0.57      0.57      0.55 235435127
weighted avg       0.59      0.56      0.56 235435127

Accuracy: 0.5569676355049601
              precision    recall  f1-score   support

           0       0.69      0.51      0.59 144486817
           1       0.45      0.63      0.52  90948310

    accuracy                           0.56 235435127
   macro avg       0.57      0.57      0.55 235435127
weighted avg       0.59      0.56      0.56 235435127

Accuracy: 0.5569676355049601

LR_1_std_scaler_0.001

              precision    recall  f1-score   support

           0       0.86      0.66      0.75  30469299
           1       0.62      0.83      0.71  19888532

    accuracy                           0.73  50357831
   macro avg       0.74      0.75      0.73  50357831
weighted avg       0.76      0.73      0.73  50357831

Accuracy: 0.7289902934858334
              precision    recall  f1-score   support

           0       0.86      0.66      0.75  30469299
           1       0.62      0.83      0.71  19888532

    accuracy                           0.73  50357831
   macro avg       0.74      0.75      0.73  50357831
weighted avg       0.76      0.73      0.73  50357831

Accuracy: 0.7289902934858334

LR_1_std_scaler_0.01

              precision    recall  f1-score   support

           0       0.86      0.68      0.76  30231476
           1       0.63      0.83      0.72  19809713

    accuracy                           0.74  50041189
   macro avg       0.74      0.76      0.74  50041189
weighted avg       0.77      0.74      0.74  50041189

Accuracy: 0.739903102622122
              precision    recall  f1-score   support

           0       0.86      0.68      0.76  30231476
           1       0.63      0.83      0.72  19809713

    accuracy                           0.74  50041189
   macro avg       0.74      0.76      0.74  50041189
weighted avg       0.77      0.74      0.74  50041189

Accuracy: 0.739903102622122

LR_1_std_scaler_0.1

              precision    recall  f1-score   support

           0       0.86      0.73      0.79  29263977
           1       0.67      0.82      0.74  18981990

    accuracy                           0.77  48245967
   macro avg       0.76      0.78      0.76  48245967
weighted avg       0.79      0.77      0.77  48245967

Accuracy: 0.7676588387170269
              precision    recall  f1-score   support

           0       0.86      0.73      0.79  29263977
           1       0.67      0.82      0.74  18981990

    accuracy                           0.77  48245967
   macro avg       0.76      0.78      0.76  48245967
weighted avg       0.79      0.77      0.77  48245967

Accuracy: 0.7676588387170269

LR_1_std_scaler_0.2

              precision    recall  f1-score   support

           0       0.87      0.75      0.81  28400293
           1       0.67      0.81      0.74  17762125

    accuracy                           0.78  46162418
   macro avg       0.77      0.78      0.77  46162418
weighted avg       0.79      0.78      0.78  46162418

Accuracy: 0.7765868113754353
              precision    recall  f1-score   support

           0       0.87      0.75      0.81  28400293
           1       0.67      0.81      0.74  17762125

    accuracy                           0.78  46162418
   macro avg       0.77      0.78      0.77  46162418
weighted avg       0.79      0.78      0.78  46162418

Accuracy: 0.7765868113754353

LR_1_std_scaler_0.4

              precision    recall  f1-score   support

           0       0.87      0.77      0.82  26121772
           1       0.66      0.80      0.72  14505950

    accuracy                           0.78  40627722
   macro avg       0.76      0.78      0.77  40627722
weighted avg       0.79      0.78      0.78  40627722

Accuracy: 0.7782466120054676
              precision    recall  f1-score   support

           0       0.87      0.77      0.82  26121772
           1       0.66      0.80      0.72  14505950

    accuracy                           0.78  40627722
   macro avg       0.76      0.78      0.77  40627722
weighted avg       0.79      0.78      0.78  40627722

Accuracy: 0.7782466120054676

LR_1_std_scaler_All

              precision    recall  f1-score   support

           0       0.69      0.51      0.59 144486817
           1       0.45      0.63      0.52  90948310

    accuracy                           0.56 235435127
   macro avg       0.57      0.57      0.55 235435127
weighted avg       0.59      0.56      0.56 235435127

Accuracy: 0.5570035307433118
              precision    recall  f1-score   support

           0       0.69      0.51      0.59 144486817
           1       0.45      0.63      0.52  90948310

    accuracy                           0.56 235435127
   macro avg       0.57      0.57      0.55 235435127
weighted avg       0.59      0.56      0.56 235435127

Accuracy: 0.5570035307433118

LR_2_0.001

              precision    recall  f1-score   support

           0       0.74      0.71      0.72  30469299
           1       0.58      0.61      0.59  19888532

    accuracy                           0.67  50357831
   macro avg       0.66      0.66      0.66  50357831
weighted avg       0.67      0.67      0.67  50357831

Accuracy: 0.6680608225560787
              precision    recall  f1-score   support

           0       0.74      0.71      0.72  30469299
           1       0.58      0.61      0.59  19888532

    accuracy                           0.67  50357831
   macro avg       0.66      0.66      0.66  50357831
weighted avg       0.67      0.67      0.67  50357831

Accuracy: 0.6680608225560787

LR_2_0.01

              precision    recall  f1-score   support

           0       0.75      0.71      0.73  30231476
           1       0.59      0.63      0.61  19809713

    accuracy                           0.68  50041189
   macro avg       0.67      0.67      0.67  50041189
weighted avg       0.68      0.68      0.68  50041189

Accuracy: 0.6778870302222435
              precision    recall  f1-score   support

           0       0.75      0.71      0.73  30231476
           1       0.59      0.63      0.61  19809713

    accuracy                           0.68  50041189
   macro avg       0.67      0.67      0.67  50041189
weighted avg       0.68      0.68      0.68  50041189

Accuracy: 0.6778870302222435

LR_2_0.1

              precision    recall  f1-score   support

           0       0.78      0.73      0.75  29263977
           1       0.62      0.68      0.65  18981990

    accuracy                           0.71  48245967
   macro avg       0.70      0.70      0.70  48245967
weighted avg       0.72      0.71      0.71  48245967

Accuracy: 0.7090260414927532
              precision    recall  f1-score   support

           0       0.78      0.73      0.75  29263977
           1       0.62      0.68      0.65  18981990

    accuracy                           0.71  48245967
   macro avg       0.70      0.70      0.70  48245967
weighted avg       0.72      0.71      0.71  48245967

Accuracy: 0.7090260414927532

LR_2_0.2

              precision    recall  f1-score   support

           0       0.80      0.74      0.77  28400293
           1       0.63      0.71      0.67  17762125

    accuracy                           0.73  46162418
   macro avg       0.72      0.72      0.72  46162418
weighted avg       0.74      0.73      0.73  46162418

Accuracy: 0.7284046299307805
              precision    recall  f1-score   support

           0       0.80      0.74      0.77  28400293
           1       0.63      0.71      0.67  17762125

    accuracy                           0.73  46162418
   macro avg       0.72      0.72      0.72  46162418
weighted avg       0.74      0.73      0.73  46162418

Accuracy: 0.7284046299307805

LR_2_0.4

              precision    recall  f1-score   support

           0       0.85      0.69      0.77  26121772
           1       0.59      0.78      0.67  14505950

    accuracy                           0.73  40627722
   macro avg       0.72      0.74      0.72  40627722
weighted avg       0.76      0.73      0.73  40627722

Accuracy: 0.726333216516545
              precision    recall  f1-score   support

           0       0.85      0.69      0.77  26121772
           1       0.59      0.78      0.67  14505950

    accuracy                           0.73  40627722
   macro avg       0.72      0.74      0.72  40627722
weighted avg       0.76      0.73      0.73  40627722

Accuracy: 0.726333216516545

LR_2_All

              precision    recall  f1-score   support

           0       0.67      0.78      0.72 144486817
           1       0.53      0.39      0.45  90948310

    accuracy                           0.63 235435127
   macro avg       0.60      0.59      0.59 235435127
weighted avg       0.62      0.63      0.62 235435127

Accuracy: 0.6319147567134279
              precision    recall  f1-score   support

           0       0.67      0.78      0.72 144486817
           1       0.53      0.39      0.45  90948310

    accuracy                           0.63 235435127
   macro avg       0.60      0.59      0.59 235435127
weighted avg       0.62      0.63      0.62 235435127

Accuracy: 0.6319147567134279

LR_2_log_0.001

              precision    recall  f1-score   support

           0       0.78      0.60      0.68  30469299
           1       0.55      0.74      0.63  19888532

    accuracy                           0.66  50357831
   macro avg       0.66      0.67      0.65  50357831
weighted avg       0.69      0.66      0.66  50357831

Accuracy: 0.6557921249626498
              precision    recall  f1-score   support

           0       0.78      0.60      0.68  30469299
           1       0.55      0.74      0.63  19888532

    accuracy                           0.66  50357831
   macro avg       0.66      0.67      0.65  50357831
weighted avg       0.69      0.66      0.66  50357831

Accuracy: 0.6557921249626498

LR_2_log_0.01

              precision    recall  f1-score   support

           0       0.80      0.61      0.69  30231476
           1       0.56      0.76      0.65  19809713

    accuracy                           0.67  50041189
   macro avg       0.68      0.69      0.67  50041189
weighted avg       0.70      0.67      0.68  50041189

Accuracy: 0.6719362523540358
              precision    recall  f1-score   support

           0       0.80      0.61      0.69  30231476
           1       0.56      0.76      0.65  19809713

    accuracy                           0.67  50041189
   macro avg       0.68      0.69      0.67  50041189
weighted avg       0.70      0.67      0.68  50041189

Accuracy: 0.6719362523540358

LR_2_log_0.1

              precision    recall  f1-score   support

           0       0.83      0.64      0.73  29263977
           1       0.59      0.80      0.68  18981990

    accuracy                           0.71  48245967
   macro avg       0.71      0.72      0.70  48245967
weighted avg       0.74      0.71      0.71  48245967

Accuracy: 0.7052470313218098
              precision    recall  f1-score   support

           0       0.83      0.64      0.73  29263977
           1       0.59      0.80      0.68  18981990

    accuracy                           0.71  48245967
   macro avg       0.71      0.72      0.70  48245967
weighted avg       0.74      0.71      0.71  48245967

Accuracy: 0.7052470313218098

LR_2_log_0.2

              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.81      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.72      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7193205087307168
              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.81      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.72      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7193205087307168

LR_2_log_0.4

              precision    recall  f1-score   support

           0       0.88      0.69      0.77  26121772
           1       0.59      0.83      0.69  14505950

    accuracy                           0.74  40627722
   macro avg       0.74      0.76      0.73  40627722
weighted avg       0.78      0.74      0.74  40627722

Accuracy: 0.7374874968377504
              precision    recall  f1-score   support

           0       0.88      0.69      0.77  26121772
           1       0.59      0.83      0.69  14505950

    accuracy                           0.74  40627722
   macro avg       0.74      0.76      0.73  40627722
weighted avg       0.78      0.74      0.74  40627722

Accuracy: 0.7374874968377504

LR_2_log_All

              precision    recall  f1-score   support

           0       0.69      0.54      0.60 144486817
           1       0.46      0.62      0.52  90948310

    accuracy                           0.57 235435127
   macro avg       0.57      0.58      0.56 235435127
weighted avg       0.60      0.57      0.57 235435127

Accuracy: 0.567944417232183
              precision    recall  f1-score   support

           0       0.69      0.54      0.60 144486817
           1       0.46      0.62      0.52  90948310

    accuracy                           0.57 235435127
   macro avg       0.57      0.58      0.56 235435127
weighted avg       0.60      0.57      0.57 235435127

Accuracy: 0.567944417232183

LR_2_mean_0.001

              precision    recall  f1-score   support

           0       0.78      0.60      0.68  30469299
           1       0.55      0.74      0.63  19888532

    accuracy                           0.66  50357831
   macro avg       0.67      0.67      0.66  50357831
weighted avg       0.69      0.66      0.66  50357831

Accuracy: 0.6570688876572146
              precision    recall  f1-score   support

           0       0.78      0.60      0.68  30469299
           1       0.55      0.74      0.63  19888532

    accuracy                           0.66  50357831
   macro avg       0.67      0.67      0.66  50357831
weighted avg       0.69      0.66      0.66  50357831

Accuracy: 0.6570688876572146

LR_2_mean_0.01

              precision    recall  f1-score   support

           0       0.80      0.60      0.69  30231476
           1       0.56      0.77      0.65  19809713

    accuracy                           0.67  50041189
   macro avg       0.68      0.69      0.67  50041189
weighted avg       0.70      0.67      0.67  50041189

Accuracy: 0.6680492943522984
              precision    recall  f1-score   support

           0       0.80      0.60      0.69  30231476
           1       0.56      0.77      0.65  19809713

    accuracy                           0.67  50041189
   macro avg       0.68      0.69      0.67  50041189
weighted avg       0.70      0.67      0.67  50041189

Accuracy: 0.6680492943522984

LR_2_mean_0.1

              precision    recall  f1-score   support

           0       0.83      0.64      0.72  29263977
           1       0.59      0.80      0.68  18981990

    accuracy                           0.70  48245967
   macro avg       0.71      0.72      0.70  48245967
weighted avg       0.74      0.70      0.71  48245967

Accuracy: 0.7039362896384687
              precision    recall  f1-score   support

           0       0.83      0.64      0.72  29263977
           1       0.59      0.80      0.68  18981990

    accuracy                           0.70  48245967
   macro avg       0.71      0.72      0.70  48245967
weighted avg       0.74      0.70      0.71  48245967

Accuracy: 0.7039362896384687

LR_2_mean_0.2

              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.82      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.73      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7202753980521558
              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.82      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.73      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7202753980521558

LR_2_mean_0.4

              precision    recall  f1-score   support

           0       0.89      0.69      0.77  26121772
           1       0.60      0.84      0.70  14505950

    accuracy                           0.74  40627722
   macro avg       0.74      0.76      0.74  40627722
weighted avg       0.78      0.74      0.75  40627722

Accuracy: 0.7423285509337688
              precision    recall  f1-score   support

           0       0.89      0.69      0.77  26121772
           1       0.60      0.84      0.70  14505950

    accuracy                           0.74  40627722
   macro avg       0.74      0.76      0.74  40627722
weighted avg       0.78      0.74      0.75  40627722

Accuracy: 0.7423285509337688

LR_2_mean_All

              precision    recall  f1-score   support

           0       0.65      0.51      0.57 144486817
           1       0.42      0.55      0.48  90948310

    accuracy                           0.53 235435127
   macro avg       0.53      0.53      0.53 235435127
weighted avg       0.56      0.53      0.54 235435127

Accuracy: 0.5302307076717571
              precision    recall  f1-score   support

           0       0.65      0.51      0.57 144486817
           1       0.42      0.55      0.48  90948310

    accuracy                           0.53 235435127
   macro avg       0.53      0.53      0.53 235435127
weighted avg       0.56      0.53      0.54 235435127

Accuracy: 0.5302307076717571

LR_3_0.001

              precision    recall  f1-score   support

           0       0.74      0.68      0.71  30469299
           1       0.56      0.64      0.60  19888532

    accuracy                           0.66  50357831
   macro avg       0.65      0.66      0.65  50357831
weighted avg       0.67      0.66      0.67  50357831

Accuracy: 0.6633904863773818
              precision    recall  f1-score   support

           0       0.74      0.68      0.71  30469299
           1       0.56      0.64      0.60  19888532

    accuracy                           0.66  50357831
   macro avg       0.65      0.66      0.65  50357831
weighted avg       0.67      0.66      0.67  50357831

Accuracy: 0.6633904863773818

LR_3_0.01

              precision    recall  f1-score   support

           0       0.75      0.71      0.73  30231476
           1       0.59      0.63      0.61  19809713

    accuracy                           0.68  50041189
   macro avg       0.67      0.67      0.67  50041189
weighted avg       0.68      0.68      0.68  50041189

Accuracy: 0.6782051881301222
              precision    recall  f1-score   support

           0       0.75      0.71      0.73  30231476
           1       0.59      0.63      0.61  19809713

    accuracy                           0.68  50041189
   macro avg       0.67      0.67      0.67  50041189
weighted avg       0.68      0.68      0.68  50041189

Accuracy: 0.6782051881301222

LR_3_0.1

              precision    recall  f1-score   support

           0       0.78      0.73      0.75  29263977
           1       0.62      0.68      0.65  18981990

    accuracy                           0.71  48245967
   macro avg       0.70      0.70      0.70  48245967
weighted avg       0.72      0.71      0.71  48245967

Accuracy: 0.7090318243595366
              precision    recall  f1-score   support

           0       0.78      0.73      0.75  29263977
           1       0.62      0.68      0.65  18981990

    accuracy                           0.71  48245967
   macro avg       0.70      0.70      0.70  48245967
weighted avg       0.72      0.71      0.71  48245967

Accuracy: 0.7090318243595366

LR_3_0.2

              precision    recall  f1-score   support

           0       0.80      0.74      0.77  28400293
           1       0.63      0.71      0.67  17762125

    accuracy                           0.73  46162418
   macro avg       0.72      0.72      0.72  46162418
weighted avg       0.74      0.73      0.73  46162418

Accuracy: 0.7283977628728201
              precision    recall  f1-score   support

           0       0.80      0.74      0.77  28400293
           1       0.63      0.71      0.67  17762125

    accuracy                           0.73  46162418
   macro avg       0.72      0.72      0.72  46162418
weighted avg       0.74      0.73      0.73  46162418

Accuracy: 0.7283977628728201

LR_3_0.4

              precision    recall  f1-score   support

           0       0.84      0.76      0.80  26121772
           1       0.63      0.74      0.68  14505950

    accuracy                           0.75  40627722
   macro avg       0.74      0.75      0.74  40627722
weighted avg       0.77      0.75      0.76  40627722

Accuracy: 0.7534139866370061
              precision    recall  f1-score   support

           0       0.84      0.76      0.80  26121772
           1       0.63      0.74      0.68  14505950

    accuracy                           0.75  40627722
   macro avg       0.74      0.75      0.74  40627722
weighted avg       0.77      0.75      0.76  40627722

Accuracy: 0.7534139866370061

LR_3_All

              precision    recall  f1-score   support

           0       0.67      0.78      0.72 144486817
           1       0.53      0.39      0.45  90948310

    accuracy                           0.63 235435127
   macro avg       0.60      0.59      0.59 235435127
weighted avg       0.62      0.63      0.62 235435127

Accuracy: 0.6319136353854283
              precision    recall  f1-score   support

           0       0.67      0.78      0.72 144486817
           1       0.53      0.39      0.45  90948310

    accuracy                           0.63 235435127
   macro avg       0.60      0.59      0.59 235435127
weighted avg       0.62      0.63      0.62 235435127

Accuracy: 0.6319136353854283

LR_3_log_0.001

              precision    recall  f1-score   support

           0       0.76      0.64      0.70  30469299
           1       0.56      0.70      0.62  19888532

    accuracy                           0.66  50357831
   macro avg       0.66      0.67      0.66  50357831
weighted avg       0.68      0.66      0.67  50357831

Accuracy: 0.6628697927835693
              precision    recall  f1-score   support

           0       0.76      0.64      0.70  30469299
           1       0.56      0.70      0.62  19888532

    accuracy                           0.66  50357831
   macro avg       0.66      0.67      0.66  50357831
weighted avg       0.68      0.66      0.67  50357831

Accuracy: 0.6628697927835693

LR_3_log_0.01

              precision    recall  f1-score   support

           0       0.79      0.62      0.70  30231476
           1       0.57      0.75      0.65  19809713

    accuracy                           0.67  50041189
   macro avg       0.68      0.69      0.67  50041189
weighted avg       0.70      0.67      0.68  50041189

Accuracy: 0.6729913232077679
              precision    recall  f1-score   support

           0       0.79      0.62      0.70  30231476
           1       0.57      0.75      0.65  19809713

    accuracy                           0.67  50041189
   macro avg       0.68      0.69      0.67  50041189
weighted avg       0.70      0.67      0.68  50041189

Accuracy: 0.6729913232077679

LR_3_log_0.1

              precision    recall  f1-score   support

           0       0.83      0.64      0.73  29263977
           1       0.59      0.80      0.68  18981990

    accuracy                           0.71  48245967
   macro avg       0.71      0.72      0.70  48245967
weighted avg       0.74      0.71      0.71  48245967

Accuracy: 0.7053253798395211
              precision    recall  f1-score   support

           0       0.83      0.64      0.73  29263977
           1       0.59      0.80      0.68  18981990

    accuracy                           0.71  48245967
   macro avg       0.71      0.72      0.70  48245967
weighted avg       0.74      0.71      0.71  48245967

Accuracy: 0.7053253798395211

LR_3_log_0.2

              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.81      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.72      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7193616677531927
              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.81      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.72      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7193616677531927

LR_3_log_0.4

              precision    recall  f1-score   support

           0       0.88      0.69      0.77  26121772
           1       0.59      0.83      0.69  14505950

    accuracy                           0.74  40627722
   macro avg       0.74      0.76      0.73  40627722
weighted avg       0.78      0.74      0.74  40627722

Accuracy: 0.7374562620075031
              precision    recall  f1-score   support

           0       0.88      0.69      0.77  26121772
           1       0.59      0.83      0.69  14505950

    accuracy                           0.74  40627722
   macro avg       0.74      0.76      0.73  40627722
weighted avg       0.78      0.74      0.74  40627722

Accuracy: 0.7374562620075031

LR_3_log_All

              precision    recall  f1-score   support

           0       0.69      0.55      0.61 144486817
           1       0.46      0.61      0.52  90948310

    accuracy                           0.57 235435127
   macro avg       0.57      0.58      0.57 235435127
weighted avg       0.60      0.57      0.58 235435127

Accuracy: 0.570340622960651
              precision    recall  f1-score   support

           0       0.69      0.55      0.61 144486817
           1       0.46      0.61      0.52  90948310

    accuracy                           0.57 235435127
   macro avg       0.57      0.58      0.57 235435127
weighted avg       0.60      0.57      0.58 235435127

Accuracy: 0.570340622960651

LR_3_mean_0.001

              precision    recall  f1-score   support

           0       0.76      0.64      0.70  30469299
           1       0.56      0.69      0.62  19888532

    accuracy                           0.66  50357831
   macro avg       0.66      0.67      0.66  50357831
weighted avg       0.68      0.66      0.67  50357831

Accuracy: 0.6620727370088676
              precision    recall  f1-score   support

           0       0.76      0.64      0.70  30469299
           1       0.56      0.69      0.62  19888532

    accuracy                           0.66  50357831
   macro avg       0.66      0.67      0.66  50357831
weighted avg       0.68      0.66      0.67  50357831

Accuracy: 0.6620727370088676

LR_3_mean_0.01

              precision    recall  f1-score   support

           0       0.80      0.61      0.69  30231476
           1       0.56      0.76      0.65  19809713

    accuracy                           0.67  50041189
   macro avg       0.68      0.68      0.67  50041189
weighted avg       0.70      0.67      0.67  50041189

Accuracy: 0.6690288873831515
              precision    recall  f1-score   support

           0       0.80      0.61      0.69  30231476
           1       0.56      0.76      0.65  19809713

    accuracy                           0.67  50041189
   macro avg       0.68      0.68      0.67  50041189
weighted avg       0.70      0.67      0.67  50041189

Accuracy: 0.6690288873831515

LR_3_mean_0.1

              precision    recall  f1-score   support

           0       0.83      0.64      0.72  29263977
           1       0.59      0.80      0.68  18981990

    accuracy                           0.70  48245967
   macro avg       0.71      0.72      0.70  48245967
weighted avg       0.74      0.70      0.71  48245967

Accuracy: 0.704026950066106
              precision    recall  f1-score   support

           0       0.83      0.64      0.72  29263977
           1       0.59      0.80      0.68  18981990

    accuracy                           0.70  48245967
   macro avg       0.71      0.72      0.70  48245967
weighted avg       0.74      0.70      0.71  48245967

Accuracy: 0.704026950066106

LR_3_mean_0.2

              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.82      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.73      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7203059207167181
              precision    recall  f1-score   support

           0       0.85      0.66      0.74  28400293
           1       0.60      0.82      0.69  17762125

    accuracy                           0.72  46162418
   macro avg       0.73      0.74      0.72  46162418
weighted avg       0.75      0.72      0.72  46162418

Accuracy: 0.7203059207167181

LR_3_mean_0.4

              precision    recall  f1-score   support

           0       0.89      0.69      0.77  26121772
           1       0.60      0.84      0.70  14505950

    accuracy                           0.74  40627722
   macro avg       0.74      0.76      0.74  40627722
weighted avg       0.78      0.74      0.75  40627722

Accuracy: 0.7423466912567729
              precision    recall  f1-score   support

           0       0.89      0.69      0.77  26121772
           1       0.60      0.84      0.70  14505950

    accuracy                           0.74  40627722
   macro avg       0.74      0.76      0.74  40627722
weighted avg       0.78      0.74      0.75  40627722

Accuracy: 0.7423466912567729

LR_3_mean_All

              precision    recall  f1-score   support

           0       0.65      0.52      0.58 144486817
           1       0.42      0.55      0.48  90948310

    accuracy                           0.53 235435127
   macro avg       0.54      0.54      0.53 235435127
weighted avg       0.56      0.53      0.54 235435127

Accuracy: 0.5348694844503811
              precision    recall  f1-score   support

           0       0.65      0.52      0.58 144486817
           1       0.42      0.55      0.48  90948310

    accuracy                           0.53 235435127
   macro avg       0.54      0.54      0.53 235435127
weighted avg       0.56      0.53      0.54 235435127

Accuracy: 0.5348694844503811

LR_4_0.001

              precision    recall  f1-score   support

           0       0.86      0.68      0.76  30469299
           1       0.63      0.83      0.72  19888532

    accuracy                           0.74  50357831
   macro avg       0.74      0.76      0.74  50357831
weighted avg       0.77      0.74      0.74  50357831

Accuracy: 0.7388880787975162
              precision    recall  f1-score   support

           0       0.86      0.68      0.76  30469299
           1       0.63      0.83      0.72  19888532

    accuracy                           0.74  50357831
   macro avg       0.74      0.76      0.74  50357831
weighted avg       0.77      0.74      0.74  50357831

Accuracy: 0.7388880787975162

LR_4_0.01

              precision    recall  f1-score   support

           0       0.86      0.69      0.76  30231476
           1       0.64      0.83      0.72  19809713

    accuracy                           0.74  50041189
   macro avg       0.75      0.76      0.74  50041189
weighted avg       0.77      0.74      0.75  50041189

Accuracy: 0.7441577777058814
              precision    recall  f1-score   support

           0       0.86      0.69      0.76  30231476
           1       0.64      0.83      0.72  19809713

    accuracy                           0.74  50041189
   macro avg       0.75      0.76      0.74  50041189
weighted avg       0.77      0.74      0.75  50041189

Accuracy: 0.7441577777058814

LR_4_0.1

              precision    recall  f1-score   support

           0       0.86      0.72      0.78  29263977
           1       0.65      0.82      0.73  18981990

    accuracy                           0.76  48245967
   macro avg       0.76      0.77      0.76  48245967
weighted avg       0.78      0.76      0.76  48245967

Accuracy: 0.7597472551436268
              precision    recall  f1-score   support

           0       0.86      0.72      0.78  29263977
           1       0.65      0.82      0.73  18981990

    accuracy                           0.76  48245967
   macro avg       0.76      0.77      0.76  48245967
weighted avg       0.78      0.76      0.76  48245967

Accuracy: 0.7597472551436268

LR_4_0.2

              precision    recall  f1-score   support

           0       0.87      0.75      0.80  28400293
           1       0.67      0.81      0.73  17762125

    accuracy                           0.77  46162418
   macro avg       0.77      0.78      0.77  46162418
weighted avg       0.79      0.77      0.78  46162418

Accuracy: 0.7725367852264584
              precision    recall  f1-score   support

           0       0.87      0.75      0.80  28400293
           1       0.67      0.81      0.73  17762125

    accuracy                           0.77  46162418
   macro avg       0.77      0.78      0.77  46162418
weighted avg       0.79      0.77      0.78  46162418

Accuracy: 0.7725367852264584

LR_4_0.4

              precision    recall  f1-score   support

           0       0.87      0.75      0.81  26121772
           1       0.64      0.80      0.72  14505950

    accuracy                           0.77  40627722
   macro avg       0.76      0.78      0.76  40627722
weighted avg       0.79      0.77      0.78  40627722

Accuracy: 0.7714900924053778
              precision    recall  f1-score   support

           0       0.87      0.75      0.81  26121772
           1       0.64      0.80      0.72  14505950

    accuracy                           0.77  40627722
   macro avg       0.76      0.78      0.76  40627722
weighted avg       0.79      0.77      0.78  40627722

Accuracy: 0.7714900924053778

LR_4_All

              precision    recall  f1-score   support

           0       0.74      0.50      0.60 144486817
           1       0.48      0.72      0.57  90948310

    accuracy                           0.59 235435127
   macro avg       0.61      0.61      0.59 235435127
weighted avg       0.64      0.59      0.59 235435127

Accuracy: 0.5873982475011046
              precision    recall  f1-score   support

           0       0.74      0.50      0.60 144486817
           1       0.48      0.72      0.57  90948310

    accuracy                           0.59 235435127
   macro avg       0.61      0.61      0.59 235435127
weighted avg       0.64      0.59      0.59 235435127

Accuracy: 0.5873982475011046

LR_4_log_0.001

              precision    recall  f1-score   support

           0       0.84      0.73      0.78  30469299
           1       0.65      0.79      0.71  19888532

    accuracy                           0.75  50357831
   macro avg       0.75      0.76      0.75  50357831
weighted avg       0.77      0.75      0.75  50357831

Accuracy: 0.7514317088041381
              precision    recall  f1-score   support

           0       0.84      0.73      0.78  30469299
           1       0.65      0.79      0.71  19888532

    accuracy                           0.75  50357831
   macro avg       0.75      0.76      0.75  50357831
weighted avg       0.77      0.75      0.75  50357831

Accuracy: 0.7514317088041381

LR_4_log_0.01

              precision    recall  f1-score   support

           0       0.85      0.73      0.78  30231476
           1       0.66      0.80      0.72  19809713

    accuracy                           0.76  50041189
   macro avg       0.75      0.76      0.75  50041189
weighted avg       0.77      0.76      0.76  50041189

Accuracy: 0.7556142600848273
              precision    recall  f1-score   support

           0       0.85      0.73      0.78  30231476
           1       0.66      0.80      0.72  19809713

    accuracy                           0.76  50041189
   macro avg       0.75      0.76      0.75  50041189
weighted avg       0.77      0.76      0.76  50041189

Accuracy: 0.7556142600848273

LR_4_log_0.1

              precision    recall  f1-score   support

           0       0.86      0.72      0.79  29263977
           1       0.66      0.82      0.73  18981990

    accuracy                           0.76  48245967
   macro avg       0.76      0.77      0.76  48245967
weighted avg       0.78      0.76      0.77  48245967

Accuracy: 0.7633229529838215
              precision    recall  f1-score   support

           0       0.86      0.72      0.79  29263977
           1       0.66      0.82      0.73  18981990

    accuracy                           0.76  48245967
   macro avg       0.76      0.77      0.76  48245967
weighted avg       0.78      0.76      0.77  48245967

Accuracy: 0.7633229529838215

LR_4_log_0.2

              precision    recall  f1-score   support

           0       0.87      0.72      0.79  28400293
           1       0.65      0.83      0.73  17762125

    accuracy                           0.76  46162418
   macro avg       0.76      0.78      0.76  46162418
weighted avg       0.79      0.76      0.77  46162418

Accuracy: 0.7624388306522418
              precision    recall  f1-score   support

           0       0.87      0.72      0.79  28400293
           1       0.65      0.83      0.73  17762125

    accuracy                           0.76  46162418
   macro avg       0.76      0.78      0.76  46162418
weighted avg       0.79      0.76      0.77  46162418

Accuracy: 0.7624388306522418

LR_4_log_0.4

              precision    recall  f1-score   support

           0       0.88      0.71      0.79  26121772
           1       0.61      0.83      0.71  14505950

    accuracy                           0.75  40627722
   macro avg       0.75      0.77      0.75  40627722
weighted avg       0.79      0.75      0.76  40627722

Accuracy: 0.7531053303948472
              precision    recall  f1-score   support

           0       0.88      0.71      0.79  26121772
           1       0.61      0.83      0.71  14505950

    accuracy                           0.75  40627722
   macro avg       0.75      0.77      0.75  40627722
weighted avg       0.79      0.75      0.76  40627722

Accuracy: 0.7531053303948472

LR_4_log_All

              precision    recall  f1-score   support

           0       0.76      0.66      0.71 144486817
           1       0.55      0.67      0.61  90948310

    accuracy                           0.66 235435127
   macro avg       0.66      0.67      0.66 235435127
weighted avg       0.68      0.66      0.67 235435127

Accuracy: 0.6641819723018647
              precision    recall  f1-score   support

           0       0.76      0.66      0.71 144486817
           1       0.55      0.67      0.61  90948310

    accuracy                           0.66 235435127
   macro avg       0.66      0.67      0.66 235435127
weighted avg       0.68      0.66      0.67 235435127

Accuracy: 0.6641819723018647

LR_4_mean_0.001

              precision    recall  f1-score   support

           0       0.87      0.71      0.78  30469299
           1       0.65      0.83      0.73  19888532

    accuracy                           0.76  50357831
   macro avg       0.76      0.77      0.76  50357831
weighted avg       0.78      0.76      0.76  50357831

Accuracy: 0.7589966096832089
              precision    recall  f1-score   support

           0       0.87      0.71      0.78  30469299
           1       0.65      0.83      0.73  19888532

    accuracy                           0.76  50357831
   macro avg       0.76      0.77      0.76  50357831
weighted avg       0.78      0.76      0.76  50357831

Accuracy: 0.7589966096832089

LR_4_mean_0.01

              precision    recall  f1-score   support

           0       0.83      0.67      0.74  30231476
           1       0.61      0.80      0.69  19809713

    accuracy                           0.72  50041189
   macro avg       0.72      0.73      0.72  50041189
weighted avg       0.75      0.72      0.72  50041189

Accuracy: 0.7192444607980838
              precision    recall  f1-score   support

           0       0.83      0.67      0.74  30231476
           1       0.61      0.80      0.69  19809713

    accuracy                           0.72  50041189
   macro avg       0.72      0.73      0.72  50041189
weighted avg       0.75      0.72      0.72  50041189

Accuracy: 0.7192444607980838

LR_4_mean_0.1

              precision    recall  f1-score   support

           0       0.86      0.66      0.74  29263977
           1       0.61      0.83      0.71  18981990

    accuracy                           0.73  48245967
   macro avg       0.74      0.75      0.72  48245967
weighted avg       0.76      0.73      0.73  48245967

Accuracy: 0.7260401890172499
              precision    recall  f1-score   support

           0       0.86      0.66      0.74  29263977
           1       0.61      0.83      0.71  18981990

    accuracy                           0.73  48245967
   macro avg       0.74      0.75      0.72  48245967
weighted avg       0.76      0.73      0.73  48245967

Accuracy: 0.7260401890172499

LR_4_mean_0.2

              precision    recall  f1-score   support

           0       0.88      0.67      0.76  28400293
           1       0.62      0.85      0.71  17762125

    accuracy                           0.74  46162418
   macro avg       0.75      0.76      0.74  46162418
weighted avg       0.78      0.74      0.74  46162418

Accuracy: 0.7395020555465703
              precision    recall  f1-score   support

           0       0.88      0.67      0.76  28400293
           1       0.62      0.85      0.71  17762125

    accuracy                           0.74  46162418
   macro avg       0.75      0.76      0.74  46162418
weighted avg       0.78      0.74      0.74  46162418

Accuracy: 0.7395020555465703

LR_4_mean_0.4

              precision    recall  f1-score   support

           0       0.89      0.68      0.77  26121772
           1       0.59      0.84      0.70  14505950

    accuracy                           0.74  40627722
   macro avg       0.74      0.76      0.73  40627722
weighted avg       0.78      0.74      0.74  40627722

Accuracy: 0.7376153159657831
              precision    recall  f1-score   support

           0       0.89      0.68      0.77  26121772
           1       0.59      0.84      0.70  14505950

    accuracy                           0.74  40627722
   macro avg       0.74      0.76      0.73  40627722
weighted avg       0.78      0.74      0.74  40627722

Accuracy: 0.7376153159657831

LR_4_mean_All

              precision    recall  f1-score   support

           0       0.72      0.55      0.62 144486817
           1       0.48      0.65      0.55  90948310

    accuracy                           0.59 235435127
   macro avg       0.60      0.60      0.59 235435127
weighted avg       0.62      0.59      0.59 235435127

Accuracy: 0.5895498168376548
              precision    recall  f1-score   support

           0       0.72      0.55      0.62 144486817
           1       0.48      0.65      0.55  90948310

    accuracy                           0.59 235435127
   macro avg       0.60      0.60      0.59 235435127
weighted avg       0.62      0.59      0.59 235435127

Accuracy: 0.5895498168376548

LR_4_robust_scaler_0.001

              precision    recall  f1-score   support

           0       0.86      0.68      0.76  30469299
           1       0.63      0.83      0.72  19888532

    accuracy                           0.74  50357831
   macro avg       0.74      0.76      0.74  50357831
weighted avg       0.77      0.74      0.74  50357831

Accuracy: 0.7388760250615242
              precision    recall  f1-score   support

           0       0.86      0.68      0.76  30469299
           1       0.63      0.83      0.72  19888532

    accuracy                           0.74  50357831
   macro avg       0.74      0.76      0.74  50357831
weighted avg       0.77      0.74      0.74  50357831

Accuracy: 0.7388760250615242

LR_4_robust_scaler_0.01

              precision    recall  f1-score   support

           0       0.86      0.69      0.76  30231476
           1       0.64      0.83      0.72  19809713

    accuracy                           0.74  50041189
   macro avg       0.75      0.76      0.74  50041189
weighted avg       0.77      0.74      0.75  50041189

Accuracy: 0.7441431897231698
              precision    recall  f1-score   support

           0       0.86      0.69      0.76  30231476
           1       0.64      0.83      0.72  19809713

    accuracy                           0.74  50041189
   macro avg       0.75      0.76      0.74  50041189
weighted avg       0.77      0.74      0.75  50041189

Accuracy: 0.7441431897231698

LR_4_robust_scaler_0.1

              precision    recall  f1-score   support

           0       0.86      0.72      0.78  29263977
           1       0.65      0.82      0.73  18981990

    accuracy                           0.76  48245967
   macro avg       0.76      0.77      0.76  48245967
weighted avg       0.78      0.76      0.76  48245967

Accuracy: 0.7597623652148997
              precision    recall  f1-score   support

           0       0.86      0.72      0.78  29263977
           1       0.65      0.82      0.73  18981990

    accuracy                           0.76  48245967
   macro avg       0.76      0.77      0.76  48245967
weighted avg       0.78      0.76      0.76  48245967

Accuracy: 0.7597623652148997

LR_4_robust_scaler_0.2

              precision    recall  f1-score   support

           0       0.87      0.75      0.80  28400293
           1       0.67      0.81      0.73  17762125

    accuracy                           0.77  46162418
   macro avg       0.77      0.78      0.77  46162418
weighted avg       0.79      0.77      0.78  46162418

Accuracy: 0.7725374784310475
              precision    recall  f1-score   support

           0       0.87      0.75      0.80  28400293
           1       0.67      0.81      0.73  17762125

    accuracy                           0.77  46162418
   macro avg       0.77      0.78      0.77  46162418
weighted avg       0.79      0.77      0.78  46162418

Accuracy: 0.7725374784310475

LR_4_robust_scaler_0.4

              precision    recall  f1-score   support

           0       0.87      0.77      0.82  26121772
           1       0.65      0.80      0.72  14505950

    accuracy                           0.78  40627722
   macro avg       0.76      0.78      0.77  40627722
weighted avg       0.79      0.78      0.78  40627722

Accuracy: 0.7776418032987427
              precision    recall  f1-score   support

           0       0.87      0.77      0.82  26121772
           1       0.65      0.80      0.72  14505950

    accuracy                           0.78  40627722
   macro avg       0.76      0.78      0.77  40627722
weighted avg       0.79      0.78      0.78  40627722

Accuracy: 0.7776418032987427

LR_4_robust_scaler_All

              precision    recall  f1-score   support

           0       0.74      0.50      0.60 144486817
           1       0.48      0.72      0.57  90948310

    accuracy                           0.59 235435127
   macro avg       0.61      0.61      0.59 235435127
weighted avg       0.64      0.59      0.59 235435127

Accuracy: 0.5873854584154725
              precision    recall  f1-score   support

           0       0.74      0.50      0.60 144486817
           1       0.48      0.72      0.57  90948310

    accuracy                           0.59 235435127
   macro avg       0.61      0.61      0.59 235435127
weighted avg       0.64      0.59      0.59 235435127

Accuracy: 0.5873854584154725

LR_4_std_scaler_0.001

              precision    recall  f1-score   support

           0       0.86      0.68      0.76  30469299
           1       0.63      0.83      0.72  19888532

    accuracy                           0.74  50357831
   macro avg       0.74      0.76      0.74  50357831
weighted avg       0.77      0.74      0.74  50357831

Accuracy: 0.7389132188795026
              precision    recall  f1-score   support

           0       0.86      0.68      0.76  30469299
           1       0.63      0.83      0.72  19888532

    accuracy                           0.74  50357831
   macro avg       0.74      0.76      0.74  50357831
weighted avg       0.77      0.74      0.74  50357831

Accuracy: 0.7389132188795026

LR_4_std_scaler_0.01

              precision    recall  f1-score   support

           0       0.86      0.69      0.76  30231476
           1       0.64      0.83      0.72  19809713

    accuracy                           0.74  50041189
   macro avg       0.75      0.76      0.74  50041189
weighted avg       0.77      0.74      0.75  50041189

Accuracy: 0.7441570183314389
              precision    recall  f1-score   support

           0       0.86      0.69      0.76  30231476
           1       0.64      0.83      0.72  19809713

    accuracy                           0.74  50041189
   macro avg       0.75      0.76      0.74  50041189
weighted avg       0.77      0.74      0.75  50041189

Accuracy: 0.7441570183314389

LR_4_std_scaler_0.1

              precision    recall  f1-score   support

           0       0.86      0.72      0.78  29263977
           1       0.65      0.82      0.73  18981990

    accuracy                           0.76  48245967
   macro avg       0.76      0.77      0.76  48245967
weighted avg       0.78      0.76      0.76  48245967

Accuracy: 0.7597217815118101
              precision    recall  f1-score   support

           0       0.86      0.72      0.78  29263977
           1       0.65      0.82      0.73  18981990

    accuracy                           0.76  48245967
   macro avg       0.76      0.77      0.76  48245967
weighted avg       0.78      0.76      0.76  48245967

Accuracy: 0.7597217815118101

LR_4_std_scaler_0.2

              precision    recall  f1-score   support

           0       0.87      0.75      0.80  28400293
           1       0.67      0.81      0.73  17762125

    accuracy                           0.77  46162418
   macro avg       0.77      0.78      0.77  46162418
weighted avg       0.79      0.77      0.78  46162418

Accuracy: 0.7726121928881629
              precision    recall  f1-score   support

           0       0.87      0.75      0.80  28400293
           1       0.67      0.81      0.73  17762125

    accuracy                           0.77  46162418
   macro avg       0.77      0.78      0.77  46162418
weighted avg       0.79      0.77      0.78  46162418

Accuracy: 0.7726121928881629

LR_4_std_scaler_0.4

              precision    recall  f1-score   support

           0       0.87      0.77      0.82  26121772
           1       0.66      0.80      0.72  14505950

    accuracy                           0.78  40627722
   macro avg       0.76      0.78      0.77  40627722
weighted avg       0.79      0.78      0.78  40627722

Accuracy: 0.7778286461643111
              precision    recall  f1-score   support

           0       0.87      0.77      0.82  26121772
           1       0.66      0.80      0.72  14505950

    accuracy                           0.78  40627722
   macro avg       0.76      0.78      0.77  40627722
weighted avg       0.79      0.78      0.78  40627722

Accuracy: 0.7778286461643111

LR_4_std_scaler_All

              precision    recall  f1-score   support

           0       0.74      0.50      0.60 144486817
           1       0.48      0.72      0.57  90948310

    accuracy                           0.59 235435127
   macro avg       0.61      0.61      0.59 235435127
weighted avg       0.64      0.59      0.59 235435127

Accuracy: 0.5872314117340698
              precision    recall  f1-score   support

           0       0.74      0.50      0.60 144486817
           1       0.48      0.72      0.57  90948310

    accuracy                           0.59 235435127
   macro avg       0.61      0.61      0.59 235435127
weighted avg       0.64      0.59      0.59 235435127

Accuracy: 0.5872314117340698

LR_5_0.001

              precision    recall  f1-score   support

           0       0.86      0.68      0.76  30469299
           1       0.63      0.83      0.72  19888532

    accuracy                           0.74  50357831
   macro avg       0.74      0.75      0.74  50357831
weighted avg       0.77      0.74      0.74  50357831

Accuracy: 0.7382249048812289
              precision    recall  f1-score   support

           0       0.86      0.68      0.76  30469299
           1       0.63      0.83      0.72  19888532

    accuracy                           0.74  50357831
   macro avg       0.74      0.75      0.74  50357831
weighted avg       0.77      0.74      0.74  50357831

Accuracy: 0.7382249048812289

LR_5_0.01

              precision    recall  f1-score   support

           0       0.86      0.69      0.76  30231476
           1       0.64      0.83      0.72  19809713

    accuracy                           0.74  50041189
   macro avg       0.75      0.76      0.74  50041189
weighted avg       0.77      0.74      0.75  50041189

Accuracy: 0.7441325384974365
              precision    recall  f1-score   support

           0       0.86      0.69      0.76  30231476
           1       0.64      0.83      0.72  19809713

    accuracy                           0.74  50041189
   macro avg       0.75      0.76      0.74  50041189
weighted avg       0.77      0.74      0.75  50041189

Accuracy: 0.7441325384974365

LR_5_0.1

              precision    recall  f1-score   support

           0       0.86      0.72      0.78  29263977
           1       0.65      0.82      0.73  18981990

    accuracy                           0.76  48245967
   macro avg       0.76      0.77      0.76  48245967
weighted avg       0.78      0.76      0.76  48245967

Accuracy: 0.7597230873204386
              precision    recall  f1-score   support

           0       0.86      0.72      0.78  29263977
           1       0.65      0.82      0.73  18981990

    accuracy                           0.76  48245967
   macro avg       0.76      0.77      0.76  48245967
weighted avg       0.78      0.76      0.76  48245967

Accuracy: 0.7597230873204386

LR_5_0.2

              precision    recall  f1-score   support

           0       0.87      0.75      0.80  28400293
           1       0.67      0.81      0.73  17762125

    accuracy                           0.77  46162418
   macro avg       0.77      0.78      0.77  46162418
weighted avg       0.79      0.77      0.78  46162418

Accuracy: 0.7726382963734698
              precision    recall  f1-score   support

           0       0.87      0.75      0.80  28400293
           1       0.67      0.81      0.73  17762125

    accuracy                           0.77  46162418
   macro avg       0.77      0.78      0.77  46162418
weighted avg       0.79      0.77      0.78  46162418

Accuracy: 0.7726382963734698

LR_5_0.4

              precision    recall  f1-score   support

           0       0.87      0.77      0.82  26121772
           1       0.66      0.80      0.72  14505950

    accuracy                           0.78  40627722
   macro avg       0.76      0.78      0.77  40627722
weighted avg       0.79      0.78      0.78  40627722

Accuracy: 0.7777381414591741
              precision    recall  f1-score   support

           0       0.87      0.77      0.82  26121772
           1       0.66      0.80      0.72  14505950

    accuracy                           0.78  40627722
   macro avg       0.76      0.78      0.77  40627722
weighted avg       0.79      0.78      0.78  40627722

Accuracy: 0.7777381414591741

LR_5_All

              precision    recall  f1-score   support

           0       0.74      0.50      0.60 144486817
           1       0.48      0.72      0.57  90948310

    accuracy                           0.59 235435127
   macro avg       0.61      0.61      0.59 235435127
weighted avg       0.64      0.59      0.59 235435127

Accuracy: 0.5873703374772958
              precision    recall  f1-score   support

           0       0.74      0.50      0.60 144486817
           1       0.48      0.72      0.57  90948310

    accuracy                           0.59 235435127
   macro avg       0.61      0.61      0.59 235435127
weighted avg       0.64      0.59      0.59 235435127

Accuracy: 0.5873703374772958

LR_5_imba_0.001

              precision    recall  f1-score   support

           0       0.81      0.78      0.79  30469299
           1       0.68      0.71      0.70  19888532

    accuracy                           0.75  50357831
   macro avg       0.74      0.75      0.74  50357831
weighted avg       0.76      0.75      0.75  50357831

Accuracy: 0.7540126182162215
              precision    recall  f1-score   support

           0       0.81      0.78      0.79  30469299
           1       0.68      0.71      0.70  19888532

    accuracy                           0.75  50357831
   macro avg       0.74      0.75      0.74  50357831
weighted avg       0.76      0.75      0.75  50357831

Accuracy: 0.7540126182162215

LR_5_imba_0.01

              precision    recall  f1-score   support

           0       0.81      0.79      0.80  30231476
           1       0.69      0.71      0.70  19809713

    accuracy                           0.76  50041189
   macro avg       0.75      0.75      0.75  50041189
weighted avg       0.76      0.76      0.76  50041189

Accuracy: 0.7600381158009655
              precision    recall  f1-score   support

           0       0.81      0.79      0.80  30231476
           1       0.69      0.71      0.70  19809713

    accuracy                           0.76  50041189
   macro avg       0.75      0.75      0.75  50041189
weighted avg       0.76      0.76      0.76  50041189

Accuracy: 0.7600381158009655

LR_5_imba_0.1

              precision    recall  f1-score   support

           0       0.81      0.82      0.81  29263977
           1       0.72      0.70      0.71  18981990

    accuracy                           0.77  48245967
   macro avg       0.76      0.76      0.76  48245967
weighted avg       0.77      0.77      0.77  48245967

Accuracy: 0.7730094828444417
              precision    recall  f1-score   support

           0       0.81      0.82      0.81  29263977
           1       0.72      0.70      0.71  18981990

    accuracy                           0.77  48245967
   macro avg       0.76      0.76      0.76  48245967
weighted avg       0.77      0.77      0.77  48245967

Accuracy: 0.7730094828444417

LR_5_imba_0.2

              precision    recall  f1-score   support

           0       0.81      0.83      0.82  28400293
           1       0.72      0.68      0.70  17762125

    accuracy                           0.77  46162418
   macro avg       0.76      0.76      0.76  46162418
weighted avg       0.77      0.77      0.77  46162418

Accuracy: 0.7742529864878396
              precision    recall  f1-score   support

           0       0.81      0.83      0.82  28400293
           1       0.72      0.68      0.70  17762125

    accuracy                           0.77  46162418
   macro avg       0.76      0.76      0.76  46162418
weighted avg       0.77      0.77      0.77  46162418

Accuracy: 0.7742529864878396

LR_5_imba_0.4

              precision    recall  f1-score   support

           0       0.81      0.86      0.83  26121772
           1       0.72      0.64      0.68  14505950

    accuracy                           0.78  40627722
   macro avg       0.76      0.75      0.76  40627722
weighted avg       0.78      0.78      0.78  40627722

Accuracy: 0.7811903163066835
              precision    recall  f1-score   support

           0       0.81      0.86      0.83  26121772
           1       0.72      0.64      0.68  14505950

    accuracy                           0.78  40627722
   macro avg       0.76      0.75      0.76  40627722
weighted avg       0.78      0.78      0.78  40627722

Accuracy: 0.7811903163066835

LR_5_imba_All

              precision    recall  f1-score   support

           0       0.64      0.93      0.76 144486817
           1       0.61      0.17      0.27  90948310

    accuracy                           0.64 235435127
   macro avg       0.62      0.55      0.51 235435127
weighted avg       0.63      0.64      0.57 235435127

Accuracy: 0.6376296685753269
              precision    recall  f1-score   support

           0       0.64      0.93      0.76 144486817
           1       0.61      0.17      0.27  90948310

    accuracy                           0.64 235435127
   macro avg       0.62      0.55      0.51 235435127
weighted avg       0.63      0.64      0.57 235435127

Accuracy: 0.6376296685753269

LR_6_0.001

              precision    recall  f1-score   support

           0       0.86      0.68      0.76  30469299
           1       0.63      0.83      0.72  19888532

    accuracy                           0.74  50357831
   macro avg       0.74      0.75      0.74  50357831
weighted avg       0.77      0.74      0.74  50357831

Accuracy: 0.7382218666248751
              precision    recall  f1-score   support

           0       0.86      0.68      0.76  30469299
           1       0.63      0.83      0.72  19888532

    accuracy                           0.74  50357831
   macro avg       0.74      0.75      0.74  50357831
weighted avg       0.77      0.74      0.74  50357831

Accuracy: 0.7382218666248751

LR_6_0.01

              precision    recall  f1-score   support

           0       0.86      0.69      0.76  30231476
           1       0.64      0.83      0.72  19809713

    accuracy                           0.74  50041189
   macro avg       0.75      0.76      0.74  50041189
weighted avg       0.77      0.74      0.75  50041189

Accuracy: 0.744151862578645
              precision    recall  f1-score   support

           0       0.86      0.69      0.76  30231476
           1       0.64      0.83      0.72  19809713

    accuracy                           0.74  50041189
   macro avg       0.75      0.76      0.74  50041189
weighted avg       0.77      0.74      0.75  50041189

Accuracy: 0.744151862578645

LR_6_0.1

              precision    recall  f1-score   support

           0       0.86      0.72      0.78  29263977
           1       0.65      0.82      0.73  18981990

    accuracy                           0.76  48245967
   macro avg       0.76      0.77      0.76  48245967
weighted avg       0.78      0.76      0.76  48245967

Accuracy: 0.759735938135513
              precision    recall  f1-score   support

           0       0.86      0.72      0.78  29263977
           1       0.65      0.82      0.73  18981990

    accuracy                           0.76  48245967
   macro avg       0.76      0.77      0.76  48245967
weighted avg       0.78      0.76      0.76  48245967

Accuracy: 0.759735938135513

LR_6_0.2

              precision    recall  f1-score   support

           0       0.87      0.75      0.80  28400293
           1       0.67      0.81      0.73  17762125

    accuracy                           0.77  46162418
   macro avg       0.77      0.78      0.77  46162418
weighted avg       0.79      0.77      0.78  46162418

Accuracy: 0.7725369801902491
              precision    recall  f1-score   support

           0       0.87      0.75      0.80  28400293
           1       0.67      0.81      0.73  17762125

    accuracy                           0.77  46162418
   macro avg       0.77      0.78      0.77  46162418
weighted avg       0.79      0.77      0.78  46162418

Accuracy: 0.7725369801902491

LR_6_0.4

              precision    recall  f1-score   support

           0       0.87      0.77      0.82  26121772
           1       0.65      0.80      0.72  14505950

    accuracy                           0.78  40627722
   macro avg       0.76      0.78      0.77  40627722
weighted avg       0.79      0.78      0.78  40627722

Accuracy: 0.7776248198213033
              precision    recall  f1-score   support

           0       0.87      0.77      0.82  26121772
           1       0.65      0.80      0.72  14505950

    accuracy                           0.78  40627722
   macro avg       0.76      0.78      0.77  40627722
weighted avg       0.79      0.78      0.78  40627722

Accuracy: 0.7776248198213033

LR_6_All

              precision    recall  f1-score   support

           0       0.74      0.50      0.60 144486817
           1       0.48      0.72      0.57  90948310

    accuracy                           0.59 235435127
   macro avg       0.61      0.61      0.59 235435127
weighted avg       0.64      0.59      0.59 235435127

Accuracy: 0.5873487434183939
              precision    recall  f1-score   support

           0       0.74      0.50      0.60 144486817
           1       0.48      0.72      0.57  90948310

    accuracy                           0.59 235435127
   macro avg       0.61      0.61      0.59 235435127
weighted avg       0.64      0.59      0.59 235435127

Accuracy: 0.5873487434183939

LR_6_imba_0.001

              precision    recall  f1-score   support

           0       0.81      0.78      0.79  30469299
           1       0.68      0.71      0.70  19888532

    accuracy                           0.75  50357831
   macro avg       0.74      0.75      0.74  50357831
weighted avg       0.76      0.75      0.75  50357831

Accuracy: 0.7540076934608244
              precision    recall  f1-score   support

           0       0.81      0.78      0.79  30469299
           1       0.68      0.71      0.70  19888532

    accuracy                           0.75  50357831
   macro avg       0.74      0.75      0.74  50357831
weighted avg       0.76      0.75      0.75  50357831

Accuracy: 0.7540076934608244

LR_6_imba_0.01

              precision    recall  f1-score   support

           0       0.81      0.79      0.80  30231476
           1       0.69      0.71      0.70  19809713

    accuracy                           0.76  50041189
   macro avg       0.75      0.75      0.75  50041189
weighted avg       0.76      0.76      0.76  50041189

Accuracy: 0.7600415130024188
              precision    recall  f1-score   support

           0       0.81      0.79      0.80  30231476
           1       0.69      0.71      0.70  19809713

    accuracy                           0.76  50041189
   macro avg       0.75      0.75      0.75  50041189
weighted avg       0.76      0.76      0.76  50041189

Accuracy: 0.7600415130024188

LR_6_imba_0.1

              precision    recall  f1-score   support

           0       0.81      0.82      0.81  29263977
           1       0.72      0.70      0.71  18981990

    accuracy                           0.77  48245967
   macro avg       0.76      0.76      0.76  48245967
weighted avg       0.77      0.77      0.77  48245967

Accuracy: 0.7730505847255585
              precision    recall  f1-score   support

           0       0.81      0.82      0.81  29263977
           1       0.72      0.70      0.71  18981990

    accuracy                           0.77  48245967
   macro avg       0.76      0.76      0.76  48245967
weighted avg       0.77      0.77      0.77  48245967

Accuracy: 0.7730505847255585

LR_6_imba_0.2

              precision    recall  f1-score   support

           0       0.81      0.83      0.82  28400293
           1       0.72      0.68      0.70  17762125

    accuracy                           0.77  46162418
   macro avg       0.76      0.76      0.76  46162418
weighted avg       0.77      0.77      0.77  46162418

Accuracy: 0.7744190523122078
              precision    recall  f1-score   support

           0       0.81      0.83      0.82  28400293
           1       0.72      0.68      0.70  17762125

    accuracy                           0.77  46162418
   macro avg       0.76      0.76      0.76  46162418
weighted avg       0.77      0.77      0.77  46162418

Accuracy: 0.7744190523122078

LR_6_imba_0.4

              precision    recall  f1-score   support

           0       0.81      0.86      0.83  26121772
           1       0.72      0.64      0.68  14505950

    accuracy                           0.78  40627722
   macro avg       0.76      0.75      0.76  40627722
weighted avg       0.78      0.78      0.78  40627722

Accuracy: 0.7814035450966215
              precision    recall  f1-score   support

           0       0.81      0.86      0.83  26121772
           1       0.72      0.64      0.68  14505950

    accuracy                           0.78  40627722
   macro avg       0.76      0.75      0.76  40627722
weighted avg       0.78      0.78      0.78  40627722

Accuracy: 0.7814035450966215

LR_6_imba_All

              precision    recall  f1-score   support

           0       0.64      0.93      0.76 144486817
           1       0.61      0.17      0.27  90948310

    accuracy                           0.64 235435127
   macro avg       0.62      0.55      0.51 235435127
weighted avg       0.63      0.64      0.57 235435127

Accuracy: 0.6376264787369643
              precision    recall  f1-score   support

           0       0.64      0.93      0.76 144486817
           1       0.61      0.17      0.27  90948310

    accuracy                           0.64 235435127
   macro avg       0.62      0.55      0.51 235435127
weighted avg       0.63      0.64      0.57 235435127

Accuracy: 0.6376264787369643

LR_7_0.001

              precision    recall  f1-score   support

           0       0.87      0.67      0.76  30469299
           1       0.63      0.84      0.72  19888532

    accuracy                           0.74  50357831
   macro avg       0.75      0.76      0.74  50357831
weighted avg       0.77      0.74      0.74  50357831

Accuracy: 0.7383836885270137
              precision    recall  f1-score   support

           0       0.87      0.67      0.76  30469299
           1       0.63      0.84      0.72  19888532

    accuracy                           0.74  50357831
   macro avg       0.75      0.76      0.74  50357831
weighted avg       0.77      0.74      0.74  50357831

Accuracy: 0.7383836885270137

LR_7_0.01

              precision    recall  f1-score   support

           0       0.87      0.69      0.77  30231476
           1       0.64      0.84      0.73  19809713

    accuracy                           0.75  50041189
   macro avg       0.75      0.77      0.75  50041189
weighted avg       0.78      0.75      0.75  50041189

Accuracy: 0.748779750217366
              precision    recall  f1-score   support

           0       0.87      0.69      0.77  30231476
           1       0.64      0.84      0.73  19809713

    accuracy                           0.75  50041189
   macro avg       0.75      0.77      0.75  50041189
weighted avg       0.78      0.75      0.75  50041189

Accuracy: 0.748779750217366

LR_7_0.1

              precision    recall  f1-score   support

           0       0.88      0.70      0.78  29263977
           1       0.65      0.86      0.74  18981990

    accuracy                           0.76  48245967
   macro avg       0.77      0.78      0.76  48245967
weighted avg       0.79      0.76      0.76  48245967

Accuracy: 0.7616268733923397
              precision    recall  f1-score   support

           0       0.88      0.70      0.78  29263977
           1       0.65      0.86      0.74  18981990

    accuracy                           0.76  48245967
   macro avg       0.77      0.78      0.76  48245967
weighted avg       0.79      0.76      0.76  48245967

Accuracy: 0.7616268733923397

LR_7_0.2

              precision    recall  f1-score   support

           0       0.89      0.70      0.78  28400293
           1       0.64      0.86      0.73  17762125

    accuracy                           0.76  46162418
   macro avg       0.76      0.78      0.76  46162418
weighted avg       0.79      0.76      0.76  46162418

Accuracy: 0.7605276006122557
              precision    recall  f1-score   support

           0       0.89      0.70      0.78  28400293
           1       0.64      0.86      0.73  17762125

    accuracy                           0.76  46162418
   macro avg       0.76      0.78      0.76  46162418
weighted avg       0.79      0.76      0.76  46162418

Accuracy: 0.7605276006122557

LR_7_0.4

              precision    recall  f1-score   support

           0       0.89      0.69      0.78  26121772
           1       0.60      0.85      0.71  14505950

    accuracy                           0.75  40627722
   macro avg       0.75      0.77      0.74  40627722
weighted avg       0.79      0.75      0.75  40627722

Accuracy: 0.7474215758392755
              precision    recall  f1-score   support

           0       0.89      0.69      0.78  26121772
           1       0.60      0.85      0.71  14505950

    accuracy                           0.75  40627722
   macro avg       0.75      0.77      0.74  40627722
weighted avg       0.79      0.75      0.75  40627722

Accuracy: 0.7474215758392755

LR_7_All

              precision    recall  f1-score   support

           0       0.71      0.32      0.44 144486817
           1       0.42      0.79      0.55  90948310

    accuracy                           0.50 235435127
   macro avg       0.57      0.56      0.50 235435127
weighted avg       0.60      0.50      0.49 235435127

Accuracy: 0.5032463571164553
              precision    recall  f1-score   support

           0       0.71      0.32      0.44 144486817
           1       0.42      0.79      0.55  90948310

    accuracy                           0.50 235435127
   macro avg       0.57      0.56      0.50 235435127
weighted avg       0.60      0.50      0.49 235435127

Accuracy: 0.5034408310702081

LR_8_0.001

              precision    recall  f1-score   support

           0       0.86      0.67      0.75  30469299
           1       0.62      0.83      0.71  19888532

    accuracy                           0.74  50357831
   macro avg       0.74      0.75      0.73  50357831
weighted avg       0.77      0.74      0.74  50357831

Accuracy: 0.73530065264328
              precision    recall  f1-score   support

           0       0.86      0.67      0.75  30469299
           1       0.62      0.83      0.71  19888532

    accuracy                           0.74  50357831
   macro avg       0.74      0.75      0.73  50357831
weighted avg       0.77      0.74      0.74  50357831

Accuracy: 0.73530065264328

LR_8_0.01

              precision    recall  f1-score   support

           0       0.87      0.69      0.77  30231476
           1       0.64      0.84      0.73  19809713

    accuracy                           0.75  50041189
   macro avg       0.75      0.77      0.75  50041189
weighted avg       0.78      0.75      0.75  50041189

Accuracy: 0.748819277655453
              precision    recall  f1-score   support

           0       0.87      0.69      0.77  30231476
           1       0.64      0.84      0.73  19809713

    accuracy                           0.75  50041189
   macro avg       0.75      0.77      0.75  50041189
weighted avg       0.78      0.75      0.75  50041189

Accuracy: 0.748819277655453

LR_8_0.1

              precision    recall  f1-score   support

           0       0.88      0.70      0.78  29263977
           1       0.65      0.86      0.74  18981990

    accuracy                           0.76  48245967
   macro avg       0.77      0.78      0.76  48245967
weighted avg       0.79      0.76      0.76  48245967

Accuracy: 0.7610481307173302
              precision    recall  f1-score   support

           0       0.88      0.70      0.78  29263977
           1       0.65      0.86      0.74  18981990

    accuracy                           0.76  48245967
   macro avg       0.77      0.78      0.76  48245967
weighted avg       0.79      0.76      0.76  48245967

Accuracy: 0.7610481307173302

LR_8_0.2

              precision    recall  f1-score   support

           0       0.89      0.70      0.78  28400293
           1       0.64      0.86      0.73  17762125

    accuracy                           0.76  46162418
   macro avg       0.76      0.78      0.76  46162418
weighted avg       0.79      0.76      0.76  46162418

Accuracy: 0.7600604023818683
              precision    recall  f1-score   support

           0       0.89      0.70      0.78  28400293
           1       0.64      0.86      0.73  17762125

    accuracy                           0.76  46162418
   macro avg       0.76      0.78      0.76  46162418
weighted avg       0.79      0.76      0.76  46162418

Accuracy: 0.7600604023818683

LR_8_0.4

              precision    recall  f1-score   support

           0       0.89      0.69      0.78  26121772
           1       0.60      0.85      0.71  14505950

    accuracy                           0.75  40627722
   macro avg       0.75      0.77      0.74  40627722
weighted avg       0.79      0.75      0.75  40627722

Accuracy: 0.7474707540826434
              precision    recall  f1-score   support

           0       0.89      0.69      0.78  26121772
           1       0.60      0.85      0.71  14505950

    accuracy                           0.75  40627722
   macro avg       0.75      0.77      0.74  40627722
weighted avg       0.79      0.75      0.75  40627722

Accuracy: 0.7474707540826434

LR_8_All

              precision    recall  f1-score   support

           0       0.71      0.34      0.46 144486817
           1       0.43      0.78      0.55  90948310

    accuracy                           0.51 235435127
   macro avg       0.57      0.56      0.50 235435127
weighted avg       0.60      0.51      0.49 235435127

Accuracy: 0.5086445150557334
              precision    recall  f1-score   support

           0       0.71      0.34      0.46 144486817
           1       0.43      0.78      0.55  90948310

    accuracy                           0.51 235435127
   macro avg       0.57      0.56      0.50 235435127
weighted avg       0.60      0.51      0.49 235435127

Accuracy: 0.5086445150557334

LR_9_0.001

              precision    recall  f1-score   support

           0       0.87      0.67      0.76  30469299
           1       0.63      0.84      0.72  19888532

    accuracy                           0.74  50357831
   macro avg       0.75      0.76      0.74  50357831
weighted avg       0.77      0.74      0.74  50357831

Accuracy: 0.7390812364416569
              precision    recall  f1-score   support

           0       0.87      0.67      0.76  30469299
           1       0.63      0.84      0.72  19888532

    accuracy                           0.74  50357831
   macro avg       0.75      0.76      0.74  50357831
weighted avg       0.77      0.74      0.74  50357831

Accuracy: 0.7390812364416569

LR_9_0.01

              precision    recall  f1-score   support

           0       0.87      0.68      0.77  30231476
           1       0.64      0.85      0.73  19809713

    accuracy                           0.75  50041189
   macro avg       0.75      0.77      0.75  50041189
weighted avg       0.78      0.75      0.75  50041189

Accuracy: 0.7483692483805691
              precision    recall  f1-score   support

           0       0.87      0.68      0.77  30231476
           1       0.64      0.85      0.73  19809713

    accuracy                           0.75  50041189
   macro avg       0.75      0.77      0.75  50041189
weighted avg       0.78      0.75      0.75  50041189

Accuracy: 0.7483692483805691

LR_9_0.1

              precision    recall  f1-score   support

           0       0.88      0.70      0.78  29263977
           1       0.65      0.86      0.74  18981990

    accuracy                           0.76  48245967
   macro avg       0.77      0.78      0.76  48245967
weighted avg       0.79      0.76      0.76  48245967

Accuracy: 0.7608571095693863
              precision    recall  f1-score   support

           0       0.88      0.70      0.78  29263977
           1       0.65      0.86      0.74  18981990

    accuracy                           0.76  48245967
   macro avg       0.77      0.78      0.76  48245967
weighted avg       0.79      0.76      0.76  48245967

Accuracy: 0.7608571095693863

LR_9_0.2

              precision    recall  f1-score   support

           0       0.89      0.70      0.78  28400293
           1       0.64      0.86      0.73  17762125

    accuracy                           0.76  46162418
   macro avg       0.76      0.78      0.76  46162418
weighted avg       0.79      0.76      0.76  46162418

Accuracy: 0.7598907622213377
              precision    recall  f1-score   support

           0       0.89      0.70      0.78  28400293
           1       0.64      0.86      0.73  17762125

    accuracy                           0.76  46162418
   macro avg       0.76      0.78      0.76  46162418
weighted avg       0.79      0.76      0.76  46162418

Accuracy: 0.7598907622213377

LR_9_0.4

              precision    recall  f1-score   support

           0       0.89      0.69      0.78  26121772
           1       0.60      0.85      0.71  14505950

    accuracy                           0.75  40627722
   macro avg       0.75      0.77      0.74  40627722
weighted avg       0.79      0.75      0.75  40627722

Accuracy: 0.7474161362037478
              precision    recall  f1-score   support

           0       0.89      0.69      0.78  26121772
           1       0.60      0.85      0.71  14505950

    accuracy                           0.75  40627722
   macro avg       0.75      0.77      0.74  40627722
weighted avg       0.79      0.75      0.75  40627722

Accuracy: 0.7474161362037478

LR_9_All

              precision    recall  f1-score   support

           0       0.71      0.34      0.46 144486817
           1       0.43      0.78      0.55  90948310

    accuracy                           0.51 235435127
   macro avg       0.57      0.56      0.50 235435127
weighted avg       0.60      0.51      0.49 235435127

Accuracy: 0.5085327262995891
              precision    recall  f1-score   support

           0       0.71      0.34      0.46 144486817
           1       0.43      0.78      0.55  90948310

    accuracy                           0.51 235435127
   macro avg       0.57      0.56      0.50 235435127
weighted avg       0.60      0.51      0.49 235435127

Accuracy: 0.5085327262995891

Model Metrics

True Positives

True Negatives

False Positives

False Negatives

Individual Workload Result

zipf_1_15

0.001

0.01

0.1

0.2

0.4

Ignore Obj Size

0.001

0.01

0.1

0.2

0.4

zipf_1_16

0.001

0.01

0.1

0.2

0.4

Ignore Obj Size

0.001

0.01

0.1

0.2

0.4

zipf_1_17

0.001

0.01

0.1

0.2

0.4

Ignore Obj Size

0.001

0.01

0.1

0.2

0.4

zipf_1_18

0.001

0.01

0.1

0.2

0.4

Ignore Obj Size

0.001

0.01

0.1

0.2

0.4

zipf_1_19

0.001

0.01

0.1

0.2

0.4

Ignore Obj Size

0.001

0.01

0.1

0.2

0.4